Supported Models

Learn about StarOps Supported Models

StarOps allows you to host and deploy several open source large language models for text generation from your own VPC. You can see our list of supported models below. If there is a model you’d like us to support please let us know by sending us feedback.

As you will be hosting these models from your own VPC, it is your responsibility to ensure that your service and/or application adheres to the acceptable use policy of each model provider, as well as any regulations governing your product and/or service in the regions you operate in.

Model NameSizeInstance TypevLLM CPU RequestvLLM CPU LimitvLLM Memory RequestvLLM Memory LimitvLLM GPU RequestvLLM GPU Limit
Gemma-3-1B-it1Bg6.xlarge2 vCPU4 vCPU8 GiB16 GiB11
Llama-3.2-3B-Instruct3Bg5.xlarge2 vCPU4 vCPU8 GiB16 GiB11
Llama-3.1-8B-Instruct8Bg5.xlarge2 vCPU4 vCPU8 GiB16 GiB11
Llama-3.1-Nemotron-Nano-8B-v18Bg5.xlarge2 vCPU4 vCPU8 GiB16 GiB11
OPT125Mg6.xlarge2 vCPU4 vCPU8 GiB16 GiB11
Phi-4-mini-instruct3.8Bg6.xlarge2 vCPU4 vCPU8 GiB16 GiB11
Qwen-2.5-7B-Instruct7Bg5.xlarge2 vCPU4 vCPU8 GiB16 GiB11