Deployment

Overview

The deployed models page in LLM Labs provides a centralized place to manage all deployed model APIs efficiently. It serves as a hub where users can oversee their deployed models, track their status, and perform necessary maintenance operations. By offering a user-friendly interface, this feature streamlines the deployment workflow, making it simple to monitor and manage models at scale.

Get started

To deploy a model, you will need to create a Sandbox first. Learn more about Sandbox.

  1. After creating and configuring your model on the Sandbox page, click the Deploy button.

  2. A dialog box will appear, allowing you to review and name your model before deployment. Once you have reviewed the details and are ready to proceed, click the Deploy model button.

  3. Once you’ve deployed the model, you will be redirected to the deployment details page. Here you can:

    1. Create API Key: Generate an API key for use when calling the API endpoint.

    2. Suspend: Temporarily pause the API endpoint. Systems using these suspended endpoints may not function until it is resumed.

    3. Delete: Delete the deployment. Deleting a deployment permanently removes its configurations and active endpoints. Any systems using these endpoints will stop functioning properly.

    4. Use in Sandbox: Choose the sandbox environment where you want to create a copy of the deployed model for iteration. After making changes, you can redeploy it to the same deployment.

    5. Review the statistics of the deployment.

  4. You can also manage your list of previously created deployments on the Deployed models page.

To learn more about how to use our Deployment API, please refer to this Deployment API page.

Last updated