Deploying a Model

This is an overview of Deploying a Model

How to initiate a model deployment

From the model list

Look for the model you are interested in using and click the deploy button - this will initiate the deployment workflow with the StarOps Agent.
You will also see any currently deployed models here with their inference endpoint

From chat

You can request to deploy a model through the chat box on the home page. Our agent may clarify the model version with you first, but once confirmed you will enter the deployment workflow.

Each deployment workflow history will be retained in StarOps - learn more here Models Page

The deployment workflow

At the start of each model deployment workflow you will be prompted to select your target environment. Your target environment is defined for your organization by the AWS Integration Settings.

Once you’ve selected your environment StarOps will begin checking your cluster dependencies.

Executioin Plan In Progress

Progress

Execution Plan Done

Done

Based on your environment, StarOps will create a deployment plan and generate all the required assets to configure your deployment and infrastructure. You can learn more about how StarOps deploys models here - How It Works.

You can see a summary of each step the agent is taking to prepare your deployment plan at the top of your conversation.

Once the agent has completed its plan you will see a summary of the resources it will create in the conversation panel.

Plan Summary

Depending on the results of the dependency check, you may see our agent include deploying the model serving framework, KServe, and required addons. At this time you will also have access to all the generated IAC assets used for your deployment.

Generated assets

The agent will generate IAC assets, in the form of Terraform files that will be used to configure your environment and launch the inference service. The files will appear in the console and you can inspect each file individually. For an overview of key resource settings you can view the config summary.

Code View

Config summary

This view provides a high level overview of critical model deployment settings contained in the deployment plan. The config summary will appear in the console once all assets and variables are finalized. It will contain: S3 bucket for the model on your VPC Key settings of your inference service, including node instance type, CPU and memory requested/limits, GPU limits, and your deployment’s scaling configurations. Once you deploy this model, the config summary will update to include your ingress and inference endpoint.

Config Summary

Deployment

Once StarOps has planned your model deployment you will have the option to deploy the model and any dependencies to your VPC. You can click the now available deployment button or simply tell the agent to deploy the model via the chat.

Once your deployment starts you will see a summary box appear of each deployment step. Note that depending on model size deployment may take a few minutes, but usually not more than 20. StatOps will run verification on the deployment to ensure all resources have been provisioned and that your cluster is ready.

Deployment Plan

Once completed you will be provided a deployment summary with your configured inference service end point.

You can see deployed models and their inference endpoint next to the models on the models page.