Datasaur
Visit our websitePricingBlogPlaygroundAPI Docs
  • Welcome to Datasaur
    • Getting started with Datasaur
  • Data Studio Projects
    • Labeling Task Types
      • Span Based
        • OCR Labeling
        • Audio Project
      • Row Based
      • Document Based
      • Bounding Box
      • Conversational
      • Mixed Labeling
      • Project Templates
        • Test Project
    • Creating a Project
      • Data Formats
      • Data Samples
      • Split Files
      • Consensus
      • Dynamic Review Capabilities
    • Pre-Labeled Project
    • Let's Get Labeling!
      • Span Based
        • Span + Line Labeling
      • Row & Document Based
      • Bounding Box Labeling
      • Conversational Labeling
      • Label Sets / Question Sets
        • Dynamic Question Set
      • Multiple Label Sets
    • Reviewing Projects
      • Review Sampling
    • Adding Documents to an Ongoing Project
    • Export Project
  • LLM Projects
    • LLM Labs Introduction
    • Sandbox
      • Direct Access LLMs
      • File Attachment
      • Conversational Prompt
    • Deployment
      • Deployment API
    • Knowledge base
      • External Object Storage
      • File Properties
    • Models
      • Amazon SageMaker JumpStart
      • Amazon Bedrock
      • Open AI
      • Azure OpenAI
      • Vertex AI
      • Custom model
      • Fine-tuning
      • LLM Comparison Table
    • Evaluation
      • Automated Evaluation
        • Multi-application evaluation
        • Custom metrics
      • Ranking (RLHF)
      • Rating
      • Performance Monitoring
    • Dataset
    • Pricing Plan
  • Workspace Management
    • Workspace
    • Role & Permission
    • Analytics
      • Inter-Annotator Agreement (IAA)
        • Cohen's Kappa Calculation
        • Krippendorff's Alpha Calculation
      • Custom Report Builder
      • Project Report
      • Evaluation Metrics
    • Activity
    • File Transformer
      • Import Transformer
      • Export Transformer
      • Upload File Transformer
      • Running File Transformer
    • Label Management
      • Label Set Management
      • Question Set Management
    • Project Management
      • Self-Assignment
        • Self-Unassign
      • Transfer Assignment Ownership
      • Reset Labeling Work
      • Mark Document as Complete
      • Project Status Workflow
        • Read-only Mode
      • Comment Feature
      • Archive Project
    • Automation
      • Action: Create Projects
  • Assisted Labeling
    • ML Assisted Labeling
      • Amazon Comprehend
      • Amazon SageMaker
      • Azure ML
      • CoreNLP NER
      • CoreNLP POS
      • Custom API
      • FewNERD
      • Google Vertex AI
      • Hugging Face
      • LLM Assisted Labeling
        • Prompt Examples
        • Custom Provider
      • LLM Labs (beta)
      • NLTK
      • Sentiment Analysis
      • spaCy
      • SparkNLP NER
      • SparkNLP POS
    • Data Programming
      • Example of Labeling Functions
      • Labeling Function Analysis
      • Inter-Annotator Agreement for Data Programming
    • Predictive Labeling
  • Assisted Review
    • Label Error Detection
  • Building Your Own Model
    • Datasaur Dinamic
      • Datasaur Dinamic with Hugging Face
      • Datasaur Dinamic with Amazon SageMaker Autopilot
  • Advanced
    • Script-Generated Question
    • Shortcuts
    • Extensions
      • Labels
      • Review
      • Document and Row Labeling
      • Bounding Box Labels
      • List of Files
      • Comments
      • Analytics
      • Dictionary
      • Search
      • Labeling Guidelines
      • Metadata
      • Grammar Checker
      • ML Assisted Labeling
      • Data Programming
      • Datasaur Dinamic
      • Predictive Labeling
      • Label Error Detection
      • LLM Sandbox
    • Tokenizers
  • Integrations
    • External Object Storage
      • AWS S3
        • With IRSA
      • Google Cloud Storage
      • Azure Blob Storage
      • Dropbox
    • SAML
      • Okta
      • Microsoft Entra ID
    • SCIM
      • Okta
      • Microsoft Entra ID
    • Webhook Notifications
      • Webhook Signature
      • Events
      • Custom Headers
    • Robosaur
      • Commands
        • Create Projects
        • Apply Project Tags
        • Export Projects
        • Generate Time Per Task Report
        • Split Document
      • Storage Options
  • API
    • Datasaur APIs
    • Credentials
    • Create Project
      • New mutation (createProject)
      • Python Script Example
    • Adding Documents
    • Labeling
      • Create Label Set
      • Add Label Sets into Existing Project
      • Get List of Label Sets in a Project
      • Add Label Set Item into Project's Label Set
      • Programmatic API Labeling
      • Inserting Span and Arrow Label into Document
    • Export Project
      • Custom Webhook
    • Get Data
      • Get List of Projects
      • Get Document Information
      • Get List of Tags
      • Get Cabinet
      • Export Team Overview
      • Check Job
    • Custom OCR
      • Importable Format
    • Custom ASR
    • Run ML-Assisted Labeling
  • Security and Compliance
    • Security and Compliance
      • 2FA
  • Compatibility & Updates
    • Common Terminology
    • Recommended Machine Specifications
    • Supported Formats
    • Supported Languages
    • Release Notes
      • Version 6
        • 6.111.0
        • 6.110.0
        • 6.109.0
        • 6.108.0
        • 6.107.0
        • 6.106.0
        • 6.105.0
        • 6.104.0
        • 6.103.0
        • 6.102.0
        • 6.101.0
        • 6.100.0
        • 6.99.0
        • 6.98.0
        • 6.97.0
        • 6.96.0
        • 6.95.0
        • 6.94.0
        • 6.93.0
        • 6.92.0
        • 6.91.0
        • 6.90.0
        • 6.89.0
        • 6.88.0
        • 6.87.0
        • 6.86.0
        • 6.85.0
        • 6.84.0
        • 6.83.0
        • 6.82.0
        • 6.81.0
        • 6.80.0
        • 6.79.0
        • 6.78.0
        • 6.77.0
        • 6.76.0
        • 6.75.0
        • 6.74.0
        • 6.73.0
        • 6.72.0
        • 6.71.0
        • 6.70.0
        • 6.69.0
        • 6.68.0
        • 6.67.0
        • 6.66.0
        • 6.65.0
        • 6.64.0
        • 6.63.0
        • 6.62.0
        • 6.61.0
        • 6.60.0
        • 6.59.0
        • 6.58.0
        • 6.57.0
        • 6.56.0
        • 6.55.0
        • 6.54.0
        • 6.53.0
        • 6.52.0
        • 6.51.0
        • 6.50.0
        • 6.49.0
        • 6.48.0
        • 6.47.0
        • 6.46.0
        • 6.45.0
        • 6.44.0
        • 6.43.0
        • 6.42.0
        • 6.41.0
        • 6.40.0
        • 6.39.0
        • 6.38.0
        • 6.37.0
        • 6.36.0
        • 6.35.0
        • 6.34.0
        • 6.33.0
        • 6.32.0
        • 6.31.0
        • 6.30.0
        • 6.29.0
        • 6.28.0
        • 6.27.0
        • 6.26.0
        • 6.25.0
        • 6.24.0
        • 6.23.0
        • 6.22.0
        • 6.21.0
        • 6.20.0
        • 6.19.0
        • 6.18.0
        • 6.17.0
        • 6.16.0
        • 6.15.0
        • 6.14.0
        • 6.13.0
        • 6.12.0
        • 6.11.0
        • 6.10.0
        • 6.9.0
        • 6.8.0
        • 6.7.0
        • 6.6.0
        • 6.5.0
        • 6.4.0
        • 6.3.0
        • 6.2.0
        • 6.1.0
        • 6.0.0
      • Version 5
        • 5.63.0
        • 5.62.0
        • 5.61.0
        • 5.60.0
  • Deployment
    • Self-Hosted
      • AWS Marketplace
        • Data Studio
        • LLM Labs
Powered by GitBook
On this page
  • Overview
  • Create fine-tuned models
  • Model management
  • Access via Sandbox
  1. LLM Projects
  2. Models

Fine-tuning

Last updated 1 month ago

Overview

LLMs in general are typically trained for generic use cases. Fine-tuning LLMs allows users to further train the model to provide more accurate answers for a specific domain or use case. This process involves providing the model with a dataset containing examples of input and output from a specific domain. LLM Labs helps simplify this process by providing a user-friendly way to fine-tune and deploy open-source models, allowing you to tailor LLMs to your exact needs.

Create fine-tuned models

This section guides you through the process of fine-tuning your models in LLM Labs.

Step 1: Set up model

  1. Navigate to the Models page.

  2. Open My models tab, and then open the Fine-tuned LLMs section. Click on the Create fine-tuned model button.

  3. Once the creation wizard is opened, you need to set up your project.

  4. Name your fine-tuned model.

    1. Select a base model that you want to fine-tune. You can select either:

      1. Pre-trained LLMs. Currently, we support:

        1. Amazon Nova Micro

        2. Amazon Nova Lite

        3. Amazon Nova Pro

        4. Amazon Titan Text G1 - Express

        5. Amazon Titan Text G1 - Lite

        6. Cohere Command

        7. Cohere Command Light

        8. Meta Llama 3.1 8B

        9. Meta Llama 3.1 70B

      2. Existing fine-tuned models

  5. Choose a dataset. You can upload either a .csv consisting of 2 columns: prompts, expected completions, or you can choose an existing dataset from the library. For the validation dataset, you have 3 options:

    1. Split from selected dataset: Datasaur will split the uploaded dataset and use it for validation data. You will need to configure the validation size using a percentage.

    2. Use new dataset: You will need to add a new dataset to use as validation.

    3. None: Choose this option if you don't want to add a validation dataset.

      1. Please note that if you select Cohere Command or Cohere Command Light as the base model, you are required to have validation data.

Step 2: Adjust hyperparameters

Next, you will need to configure and adjust the hyperparameters for your fine-tuning project.

The fundamental hyperparameters are:

Models
Hyperparameters

Amazon Titan Text G1 - Express

Epochs, and Learning rate.

Amazon Titan Text G1 - Lite

Epochs, and Learning rate.

Cohere Command

Epochs, and Learning rate.

Cohere Command Light

Epochs, and Learning rate.

Meta Llama 2 13B

Epochs, and Learning rate.

Meta Llama 2 70B

Epochs, and Learning rate.

In addition to the fundamental hyperparameters, there are advanced hyperparameters with preset default values. These hyperparameters are always applied, but you can adjust them for further hyperparameter fine-tuning if desired. The advanced hyperparameters include:

Models
Hyperparameters

Amazon Titan Text G1 - Express

Batch size, and Learning rate warmup steps

Amazon Titan Text G1 - Lite

Batch size, and Learning rate warmup steps

Cohere Command

Batch size, Early stopping threshold, and Early stopping patience.

Cohere Command Light

Batch size, Early stopping threshold, and Early stopping patience.

Meta Llama 2 13B

Batch size.

Meta Llama 2 70B

Batch size.

Step 3: Review job

  1. The last step of the creation wizard is to review your fine-tuning job before you start the process. In this step, you will need to review all the configurations that you have chosen. The important part of this page is to review the cost that it will take to start the fine-tuning process.

    You can also view the predicted cost by clicking the View total predicted cost button on the Costs section.

    It will calculate and show you the total predicted cost for starting the fine-tuning process.

Please note that this is just a cost prediction. The final cost may be higher or lower, as each model has its own tokenizer.

  1. Once you have reviewed the configuration, you will need to check the acknowledgement checkbox.

  2. Lastly, click the Start fine-tuning job button to start the fine-tuning process.

  3. The training process has now started.

It will take several hours for the training process to be completed. Datasaur will notify you by email when the training process is complete.

  1. Once the training process is complete, your model will be available in the My models section.

Model management

Model status

There are 7 possible statuses for the fine-tuned model.

  1. Training: The model is currently being trained on your dataset. This status indicates that the training process is in progress, and the model is learning from your data.

  2. Training failed: The model training process has failed due to an error. This status indicates that the training process was interrupted, and you may need to investigate and resolve the issue.

  3. Stopping training: The model training process is being stopped. This status indicates that someone has chosen to stop the training.

  4. Training stopped: The model training process has been stopped. This status indicates that the training process has been successfully stopped, and you can’t continue the training once it stopped.

  5. Not deployed: The model has been trained but has not yet been deployed for use. You can deploy the model to use it in Sandbox.

  6. Deploying: The model is being deployed for use. This status indicates that the deployment process is in progress, and the model will soon be available for use in Sandbox.

  7. Deployed: The model has been successfully deployed. This status indicates that the model is now available for use in Sandbox, and you can start using it to generate predictions or responses.

Learn more about deploying or undeploying models.

View model details

To view the model details, click the three-dots button in the top right corner of the model card.

Choose the View details menu, and the details of the fine-tuned model will be shown.

Here you can view the dataset, validation dataset, models used, hyperparameter configurations, the creator, and storage cost information.

Delete fine-tune models

To delete the fine-tune model, click the three-dots button in the top right corner of the model card.

Choose the Delete option.

Acknowledge that you understand the impact of deleting the fine-tuned model by checking the acknowledgement checkbox.

After that, confirm the delete process by clicking the Delete model button, and your model will be deleted.

Deploy and undeploy model

To deploy a fine-tuned model:

  1. Navigate to the My models tab, and expand the Fine-tuned LLMs section.

  2. Click the Deploy model button to start the deployment.

  3. Once you click the Deploy model button, you can specify the auto undeploy schedule based on your needs.

  4. Click the Deploy model button to start the deployment process.

To undeploy a model:

  1. Navigate to the My models tab, and expand the Fine-tuned LLMs section.

  2. Click the three-dots button in the right corner of the model card.

  3. Choose the Undeploy option.

  4. Confirm the undeployment by clicking the Undeploy button, and your model will be undeployed. You will no longer be charged for the hourly cost.

Access via Sandbox

Learn more on how to create the Dataset .

Once deployment is complete, your model will be available to use for experiment in Sandbox. .

Once the fine-tuned model is deployed, it will be available in the Sandbox for further experimentation and testing. This allows you to integrate and test the specialized model within your specific applications. .

here
Learn more about Sandbox
Learn more about Sandbox
My models page
Set up model
Split from selected dataset
Use new dataset
Adjust hyperparameters
Review job
Costs section
Total predicted costs
Training process started
My models page
Training status
Training failed
Stopping training
Training stopped
Not deployed
Deploying
Deployed
Model card
View details menu
View details
Delete menu
Delete model dialog
Delete model success
Deploy model
Deploy model dialog
Deploying process
Model deployed
Undeploy model
Undeploy model dialog
Access model via Sandbox