Skip to content
Verda Documentation
Job Orchestrators
Search
Verda Documentation
Welcome to Verda
Welcome to Verda
Overview
Locations and Sustainability
Pricing and Billing
Team Projects
Release Notes
Release Notes
overview
Verda API changes
Support
CPU and GPU Instances
CPU and GPU Instances
Set up a CPU or GPU instance
Securing Your Instance
Shutdown and Delete
Adding a New User
Managing SSH Keys
Connecting to Your Server
Accessing JupyterLab
Connecting to Jupyter notebook with VS Code
Remote desktop access
Troubleshooting SSH Connection Issues
Confidential Compute
Tips and tricks
Tips and tricks
Setting environment variables on startup
Best ways to copy files between two block devices
Clusters
Clusters
Instant Clusters
Instant Clusters
Overview
Deploying an Instant Cluster
Job Orchestrators
Job Orchestrators
Overview
Kubernetes
Slurm
Environments
Containers
Monitoring
Validation
Good to know
Local Users
Local Users
Overview
Use Case: workshop
Tutorial: deploying vllm inference on instant cluster using ray
Customized GPU clusters
Storage
Storage
Block Volumes
Block Volumes
Overview
Attaching a block volume
Resizing a block volume
Cloning a block volume
Shared Filesystems (SFS)
Shared Filesystems (SFS)
Overview
Creating a shared filesystem
Editing share settings
Mounting a shared filesystem
Shared filesystem for a cluster
Deleting storage
Container registry
Container registry
Overview
Quickstart
Tag immutability rules
Tag retention rules & retention runs
Tag rules syntax
Containers
Containers
Overview
Container Registries
Scaling and health-checks
Batching and Streaming
Async Inference
Storage
Batch Jobs
Tutorials
Tutorials
Overview
Quick: Deploy with vLLM
Quick: Migrate from Runpod
Quick: Deploying GPT-OSS 120B (Ollama) on Serverless Containers
In-Depth: Deploy with TGI
In-Depth: Deploy with SGLang
In-Depth: Deploy with vLLM
In-Depth: Deploy with Replicate Cog
In-Depth: Asynchronous Inference Requests with Whisper
Tutorial: How to Publish Your First Docker Image to Docker Hub
Inference
Inference
Overview
Getting Started
Authorization
Image Models
Image Models
Overview
FLUX.2 [klein]
FLUX.2
FLUX.1 Kontext [dev]
FLUX.1 Kontext [pro]
FLUX.1 Kontext [max]
FLUX.1 Krea [dev]
FLUX.1 [dev]
Audio Models
Audio Models
Overview
Whisper
Pricing and Billing
Infrastructure as Code
Infrastructure as Code
Overview
Terraform
Terraform
Overview
Getting Started
Authentication
Provider Configuration
Compute – Instances
Compute – SSH Keys
Compute – Startup Scripts
Storage – Volumes
Containers – Containers
Containers – Serverless Jobs
OpenTofu
OpenTofu
Overview
Getting Started
Using Verda with OpenTofu
Migration from Terraform
Verda CLI
Verda CLI
Overview
Getting Started
Instances
Templates
Storage
SSH Keys & Startup Scripts
Cost & Status
Skills
MCP Server
Resources
Resources
Resources Overview
Services Overview
Shared Responsibility Model
Verda API
Python SDK
Get Free Compute Credits
Get Free Compute Credits
Overview
How to Receive Credits
How to Redeem Credits
Integrations
Integrations
dstack
Job Orchestrators
¶
Kubernetes
SLURM