Datasaur
Visit our websitePricingBlogPlaygroundAPI Docs
  • Welcome to Datasaur
    • Getting started with Datasaur
  • Data Studio Projects
    • Labeling Task Types
      • Span Based
        • OCR Labeling
        • Audio Project
      • Row Based
      • Document Based
      • Bounding Box
      • Conversational
      • Mixed Labeling
      • Project Templates
        • Test Project
    • Creating a Project
      • Data Formats
      • Data Samples
      • Split Files
      • Consensus
      • Dynamic Review Capabilities
    • Pre-Labeled Project
    • Let's Get Labeling!
      • Span Based
        • Span + Line Labeling
      • Row & Document Based
      • Bounding Box Labeling
      • Conversational Labeling
      • Label Sets / Question Sets
        • Dynamic Question Set
      • Multiple Label Sets
    • Reviewing Projects
      • Review Sampling
    • Adding Documents to an Ongoing Project
    • Export Project
  • LLM Projects
    • LLM Labs Introduction
    • Sandbox
      • Direct Access LLMs
      • File Attachment
      • Conversational Prompt
    • Deployment
      • Deployment API
    • Knowledge base
      • External Object Storage
      • File Properties
    • Models
      • Amazon SageMaker JumpStart
      • Amazon Bedrock
      • Open AI
      • Azure OpenAI
      • Vertex AI
      • Custom model
      • Fine-tuning
      • LLM Comparison Table
    • Evaluation
      • Automated Evaluation
        • Multi-application evaluation
        • Custom metrics
      • Ranking (RLHF)
      • Rating
      • Performance Monitoring
    • Dataset
    • Pricing Plan
  • Workspace Management
    • Workspace
    • Role & Permission
    • Analytics
      • Inter-Annotator Agreement (IAA)
        • Cohen's Kappa Calculation
        • Krippendorff's Alpha Calculation
      • Custom Report Builder
      • Project Report
      • Evaluation Metrics
    • Activity
    • File Transformer
      • Import Transformer
      • Export Transformer
      • Upload File Transformer
      • Running File Transformer
    • Label Management
      • Label Set Management
      • Question Set Management
    • Project Management
      • Self-Assignment
        • Self-Unassign
      • Transfer Assignment Ownership
      • Reset Labeling Work
      • Mark Document as Complete
      • Project Status Workflow
        • Read-only Mode
      • Comment Feature
      • Archive Project
    • Automation
      • Action: Create Projects
  • Assisted Labeling
    • ML Assisted Labeling
      • Amazon Comprehend
      • Amazon SageMaker
      • Azure ML
      • CoreNLP NER
      • CoreNLP POS
      • Custom API
      • FewNERD
      • Google Vertex AI
      • Hugging Face
      • LLM Assisted Labeling
        • Prompt Examples
        • Custom Provider
      • LLM Labs (beta)
      • NLTK
      • Sentiment Analysis
      • spaCy
      • SparkNLP NER
      • SparkNLP POS
    • Data Programming
      • Example of Labeling Functions
      • Labeling Function Analysis
      • Inter-Annotator Agreement for Data Programming
    • Predictive Labeling
  • Assisted Review
    • Label Error Detection
  • Building Your Own Model
    • Datasaur Dinamic
      • Datasaur Dinamic with Hugging Face
      • Datasaur Dinamic with Amazon SageMaker Autopilot
  • Advanced
    • Script-Generated Question
    • Shortcuts
    • Extensions
      • Labels
      • Review
      • Document and Row Labeling
      • Bounding Box Labels
      • List of Files
      • Comments
      • Analytics
      • Dictionary
      • Search
      • Labeling Guidelines
      • Metadata
      • Grammar Checker
      • ML Assisted Labeling
      • Data Programming
      • Datasaur Dinamic
      • Predictive Labeling
      • Label Error Detection
      • LLM Sandbox
    • Tokenizers
  • Integrations
    • External Object Storage
      • AWS S3
        • With IRSA
      • Google Cloud Storage
      • Azure Blob Storage
    • SAML
      • Okta
      • Microsoft Entra ID
    • SCIM
      • Okta
      • Microsoft Entra ID
    • Webhook Notifications
      • Webhook Signature
      • Events
      • Custom Headers
    • Robosaur
      • Commands
        • Create Projects
        • Apply Project Tags
        • Export Projects
        • Generate Time Per Task Report
        • Split Document
      • Storage Options
  • API
    • Datasaur APIs
    • Credentials
    • Create Project
      • New mutation (createProject)
      • Python Script Example
    • Adding Documents
    • Labeling
      • Create Label Set
      • Add Label Sets into Existing Project
      • Get List of Label Sets in a Project
      • Add Label Set Item into Project's Label Set
      • Programmatic API Labeling
      • Inserting Span and Arrow Label into Document
    • Export Project
      • Custom Webhook
    • Get Data
      • Get List of Projects
      • Get Document Information
      • Get List of Tags
      • Get Cabinet
      • Export Team Overview
      • Check Job
    • Custom OCR
      • Importable Format
    • Custom ASR
    • Run ML-Assisted Labeling
  • Security and Compliance
    • Security and Compliance
      • 2FA
  • Compatibility & Updates
    • Common Terminology
    • Recommended Machine Specifications
    • Supported Formats
    • Supported Languages
    • Release Notes
      • Version 6
        • 6.111.0
        • 6.110.0
        • 6.109.0
        • 6.108.0
        • 6.107.0
        • 6.106.0
        • 6.105.0
        • 6.104.0
        • 6.103.0
        • 6.102.0
        • 6.101.0
        • 6.100.0
        • 6.99.0
        • 6.98.0
        • 6.97.0
        • 6.96.0
        • 6.95.0
        • 6.94.0
        • 6.93.0
        • 6.92.0
        • 6.91.0
        • 6.90.0
        • 6.89.0
        • 6.88.0
        • 6.87.0
        • 6.86.0
        • 6.85.0
        • 6.84.0
        • 6.83.0
        • 6.82.0
        • 6.81.0
        • 6.80.0
        • 6.79.0
        • 6.78.0
        • 6.77.0
        • 6.76.0
        • 6.75.0
        • 6.74.0
        • 6.73.0
        • 6.72.0
        • 6.71.0
        • 6.70.0
        • 6.69.0
        • 6.68.0
        • 6.67.0
        • 6.66.0
        • 6.65.0
        • 6.64.0
        • 6.63.0
        • 6.62.0
        • 6.61.0
        • 6.60.0
        • 6.59.0
        • 6.58.0
        • 6.57.0
        • 6.56.0
        • 6.55.0
        • 6.54.0
        • 6.53.0
        • 6.52.0
        • 6.51.0
        • 6.50.0
        • 6.49.0
        • 6.48.0
        • 6.47.0
        • 6.46.0
        • 6.45.0
        • 6.44.0
        • 6.43.0
        • 6.42.0
        • 6.41.0
        • 6.40.0
        • 6.39.0
        • 6.38.0
        • 6.37.0
        • 6.36.0
        • 6.35.0
        • 6.34.0
        • 6.33.0
        • 6.32.0
        • 6.31.0
        • 6.30.0
        • 6.29.0
        • 6.28.0
        • 6.27.0
        • 6.26.0
        • 6.25.0
        • 6.24.0
        • 6.23.0
        • 6.22.0
        • 6.21.0
        • 6.20.0
        • 6.19.0
        • 6.18.0
        • 6.17.0
        • 6.16.0
        • 6.15.0
        • 6.14.0
        • 6.13.0
        • 6.12.0
        • 6.11.0
        • 6.10.0
        • 6.9.0
        • 6.8.0
        • 6.7.0
        • 6.6.0
        • 6.5.0
        • 6.4.0
        • 6.3.0
        • 6.2.0
        • 6.1.0
        • 6.0.0
      • Version 5
        • 5.63.0
        • 5.62.0
        • 5.61.0
        • 5.60.0
  • Deployment
    • Self-Hosted
      • AWS Marketplace
        • Data Studio
        • LLM Labs
Powered by GitBook
On this page
  • Overview
  • Supported Providers
  • Azure OpenAI
  • OpenAI
  • Amazon Bedrock
  • Vertex AI
  • Azure AI
  • Hugging Face
  • Getting Started with Direct Access LLM
  1. LLM Projects
  2. Sandbox

Direct Access LLMs

Last updated 2 months ago

Overview

Datasaur offers Direct Access LLMs, a new feature that allows users to instantly access and call the most popular Large Language Models (LLMs) within the platform. This feature eliminates the need for complex API key setup and multi-cloud configurations. Additionally, users can cut wait lines and immediately access the latest state of the art models.

Supported Providers

Datasaur's Direct Access LLM feature currently supports Azure OpenAI, OpenAI, Amazon Bedrock and Google Vertex, each offering a unique set of LLM models. Below, we'll delve into the details of each provider and the models they offer.

Azure OpenAI

With Azure OpenAI, users can utilize the following models:

  • gpt-4o: A highly advanced model boasting an expansive knowledge base for richer and more comprehensive responses.

  • gpt-4-32k: A variant of the gpt-4 model, with greater capacity to handle longer inputs.

  • gpt-4 turbo: A high-performance model optimized for speed and efficiency.

  • gpt-4: A powerful model offering advanced language understanding and generation capabilities.

  • gpt-35-turbo-16k: A variant of the gpt-35-turbo model, with greater capacity to handle longer inputs.

  • gpt-35-turbo: A fast and efficient model ideal for applications requiring rapid response times.

OpenAI

With OpenAI, users can utilize the following model:

  • o1-mini-2024-09-12: A compact and efficient model well-suited for tasks that require fast inference and lower computational resources. It excels in short-form text generation, question answering, and text summarization.

  • o1-preview-2024-09-12: A preview model offering advanced capabilities and access to the latest developments in OpenAI's LLM technology. This model is ideal for exploring cutting-edge language processing tasks and experimenting with potential future functionalities.

  • gpt-4o-mini: A streamlined and efficient version of the advanced gpt-4o model, designed to deliver rich responses while requiring less computational power.

  • gpt-4o: A highly advanced model boasting an expansive knowledge base for richer and more comprehensive responses.

  • gpt-4 turbo: A high-performance model optimized for speed and efficiency.

  • gpt-4: A powerful model offering advanced language understanding and generation capabilities.

  • gpt-35-turbo-16k: A variant of the gpt-35-turbo model, with greater capacity to handle longer inputs.

  • gpt-35-turbo: A fast and efficient model ideal for applications requiring rapid response times.

Amazon Bedrock

With Amazon Bedrock, Datasaur is able to provide several Open Source Models, such as:

  • Claude 3.5 Sonnet: An enhanced version of Claude 3 Sonnet, with updated knowledge and improved reasoning capabilities.

  • Claude 3 Sonnet: A more verbose Claude model, offering deeper analysis and extended conversations.

  • Claude 3 Opus: The most comprehensive Claude model, providing in-depth expertise across a wide range of subjects.

  • Claude 3 Haiku: A concise and efficient AI assistant, perfect for brief, focused interactions.

  • Claude 2.1: An updated version of Claude 2.0, featuring refinements in language understanding and generation.

  • Claude 2.0: An upgraded Claude model with expanded knowledge and improved conversational abilities.

  • Claude Instant: A rapid-response AI assistant for quick, concise interactions.

  • Meta Llama 3 70b Instruct: A variant of the Meta Llama 3 8b Instruct model, with increased capacity and performance.

  • Meta Llama 3 8b Instruct: A newer model optimized for instruction-following tasks, offering high accuracy and reliability.

  • Meta Llama 2 Chat 70B: A variant of the Meta Llama 2 Chat 13B model, with increased capacity and performance.

  • Meta Llama 2 Chat 13B: A highly advanced model designed for conversational AI applications.

  • Mistral Large: A more expansive version of Mistral, offering deeper knowledge and more nuanced interactions.

  • Mixtral 8x7B Instruct: An advanced instruction-following model combining multiple expert systems for enhanced performance.

  • Mistral 7B Instruct: A compact yet powerful model designed for following instructions with precision.

  • Mistral Small: A nimble AI assistant optimized for quick responses and everyday tasks.

  • Command R+: The most advanced Command model, featuring superior problem-solving and creative abilities.

  • Command R: An enhanced version of Command, with improved reasoning and analytical skills.

  • Command: A versatile AI assistant balancing speed and capability for various applications.

  • Command Light: A streamlined AI model for efficient, straightforward task completion.

  • Amazon Titan Text Premier: Amazon's most advanced text AI, offering sophisticated language understanding and generation.

  • Amazon Titan Text Express: A mid-range AI assistant balancing efficiency and capability for various text-based applications.

  • Amazon Titan Text Lite: A lightweight AI model for basic text processing and generation tasks.

Vertex AI

With Vertex AI, users can utilize the following models:

  • Gemini 1.5 Pro: A high-performance model offering advanced language understanding and generation capabilities.

  • Gemini 1.5 Flash: A variant of the Gemini 1.0 Pro model, optimized for speed and efficiency.

  • Gemini 1.0 Pro: A highly advanced model offering exceptional language understanding and generation capabilities.

Azure AI

With Azure AI, users can utilize the following models:

  • Meta-Llama-3-1-405B-Instruct: A massive 405 billion parameter version of Meta-Llama hosted on Azure AI, specifically optimized for following instructions and demonstrating exceptional proficiency in complex language understanding and generation.

  • Meta-Llama-3-1-70B-Instruct: A powerful 70 billion parameter model hosted on Azure AI, fine-tuned for instruction following. This model offers a balance between scale and efficiency, making it well-suited for a wide range of LLM tasks.

  • Meta-Llama-3-1-8B-Instruct: An efficient 8 billion parameter version of Meta-Llama on Azure AI, suitable for tasks where resource efficiency is crucial while maintaining respectable language processing capabilities.

Hugging Face

With Hugging Face, users can utilize the following models:

  • Meta-Llama-3.1-70B-Instruct: A powerful 70 billion parameter model from Meta fine-tuned for following instructions. This model excels at complex language tasks, text generation, question answering, and code generation.

  • Meta-Llama-3.1-8B-Instruct: A smaller 8 billion parameter version of Meta-Llama, offering a good balance between performance and efficiency. It is suitable for tasks where resource constraints are a factor while still maintaining good language understanding and generation capabilities.

  • Mistral-7B-Instruct-v0.1: The first iteration of the Mistral-7B model fine-tuned for instruction following. This model excels in code generation, reasoning tasks, and understanding complex instructions.

  • Mistral-7B-Instruct-v0.2: An improved version of the Mistral-7B-Instruct model with enhanced instruction-following capabilities and performance.

  • Mistral-7B-Instruct-v0.3: The latest iteration of the Mistral-7B-Instruct model, further refined for better accuracy, coherence, and instruction adherence in text generation tasks.

  • Mistral-Nemo-Instruct-2407: A specialized version of the Mistral model trained on a massive dataset for improved code generation and technical language understanding. This model is particularly useful for tasks involving code-related data or highly technical language.

  • Mixtral-8x7B-Instruct-v0.1: A powerful model combining eight smaller 7 billion parameter models for enhanced performance and capabilities. This model showcases advancements in mixture-of-experts architecture, demonstrating strong performance in various natural language processing tasks.

Getting Started with Direct Access LLM

Once the Sandbox is created, you just have to choose your desired provider by clicking the models button.

To get started with Direct Access LLM, simply navigate to the Datasaur LLM Labs platform and create a new .

Sandbox
Azure OpenAI
OpenAI
Amazon Bedrock
Vertex AI
Sandbox
Application configuration