Tutorials¶ On this section you'll find tutorials how to deploy a selection of models using Verda serverless containers. Quick Tutorial: Deploy with vLLM In-Depth: Deploy with Text Generation Inference In-Depth: Deploy with SGLang In-Depth: Deploy with vLLM In-Depth: Deploy with Replicate Cog In-Depth: Asynchronous Inference Requests with Whisper