LLM Assisted Labeling
Last updated
Last updated
Datasaur ML Assisted Labeling with Large Language Model supports users to integrate the labeling process to be assisted labeled with recents top notch model from OpenAI, Azure OpenAI, Anthropic, Gemini and Cohere. This will make labeling easier and general for all use cases!
OpenAI
gpt-3.5-turbo, gpt-4, gpt-4-turbo, gpt-4o, gpt-4o-mini
Azure Open AI
Anthropic
claude-2, claude-2.1, claude-3-haiku-20240307, claude-3-opus-20240229, claude-3-sonnet-20240229, claude-3-5-sonnet-20240620
Gemini
gemini-pro, gemini-1.5-flash, gemini-1.5-pro
Cohere
command-light, command-r, command-r-plus
After successfully creating the project, you need to activate the ML-assisted labeling extension and select LLM Assisted Labeling as the provider. Once you have chosen LLM Assisted Labeling, you can access several fields under the extension. These fields include:
LLM provider: You can choose from the variety of LLM that Datasaur support.
Target text: Define your text column(s) that is going to be treated as input and prompt context.
Target question: Select your question to be answered.
System prompt: Sets the behavior and context for the language model.
User prompt: User definition of a task to be completed in a specific labeling workflow.
API key: The LLM Provider secret key
API version: The API Version from your Azure OpenAI.
API base URL: The base URL for your Azure OpenAI API model.
Model deployment: The deployment model name from Azure OpenAI.
Advanced Settings
Top P: Limits predictions to the smallest set with a cumulative probability of P.
Temperature: Controls randomness; lower values make responses more predictable.
Maximum tokens: Limits the length of the generated response.
Model Name: The specific version of the language model.
For guidance, you can refer to our prompt examples: Row-Based and Span-Based.
Once all fields have been filled, you can predict the label by clicking “Predict label” then you will see the assisted labeling recommendation from your prompts and settings.
If you are experiencing the 429 error, the limitation came from the LLM Provider package. Please take a look at your current usage of LLM Provider API.