On this section you'll find tutorials how to deploy a selection of models using Verda serverless containers.
Quick: Deploy with vLLM
Quick: Migrate from Runpod
Quick: Deploying GPT-OSS 120B (Ollama)
In-Depth: Deploy with TGI
In-Depth: Deploy with SGLang
In-Depth: Deploy with vLLM
In-Depth: Deploy with Replicate Cog
In-Depth: Async inference with Whisper
Tutorial: Publish an image to Docker Hub
Last updated 10 days ago
Was this helpful?