On this section you'll find tutorials how to deploy a selection of models using Verda serverless containers.
Quick Tutorial: Deploy with vLLM
In-Depth: Deploy with Text Generation Inference
In-Depth: Deploy with SGLang
In-Depth: Deploy with vLLM
In-Depth: Deploy with Replicate Cog
In-Depth: Asynchronous Inference Requests with Whisper
Last updated 1 month ago
Was this helpful?