Container orchestration for AI teams

dstack is an open-source container orchestrator that simplifies workload orchestration and drives GPU utilization for ML teams. It works with any GPU cloud, on-prem cluster, or accelerated hardware.

Unified compute layer for AI

dstack redefines the container orchestration layer for AI workloads, tailoring the developer experience for ML teams while keeping it open and vendor-agnostic.

dstack natively integrates with leading GPU clouds and on-prem clusters, regardless of hardware vendor.

Backends

dstack manages instances and clusters via native GPU cloud integration for efficient provisioning with its built-in scheduler. It also runs smoothly on Kubernetes if that fits your environment.

In both cases, dstack acts as an enhanced control plane, making AI compute orchestration more efficient.

Learn more

SSH fleets

Bare-metal and manually provisioned clusters can be connected to dstack using SSH fleets.

Learn more

Dev environments

Before running training jobs or deploying model endpoints, ML engineers often experiment with their code in a desktop IDE while using cloud or on-prem GPU machines. Dev environments simplify and streamline this process.

Learn more

Tasks

Tasks simplify the process of scheduling jobs on either optimized clusters or individual instances. They can be used for pre-training or fine-tuning models, as well as for running any AI or data workloads that require efficient GPU utilization.

Learn more

Services

With dstack, you can easily deploy any model as a secure, auto-scaling OpenAI-compatible endpoint, all while using your custom code, Docker image, and serving framework.

Learn more

Loved by world-class ML teams

Wah Loon Keng

Sr. AI Engineer @Electronic Arts

With dstack, AI researchers at EA can spin up and scale experiments without touching infrastructure. It supports everything from quick prototyping to multi-node training on any cloud.

Aleksandr Movchan

ML Engineer @Mobius Labs

Thanks to dstack, my team can quickly tap into affordable GPUs and streamline our workflows from testing and development to full-scale application deployment.

Alvaro Bartolome

ML Engineer @Argilla

With dstack it's incredibly easy to define a configuration within a repository and run it without worrying about GPU availability. It lets you focus on data and your research.

Park Chansung

ML Researcher @ETRI

Thanks to dstack, I can effortlessly access the top GPU options across different clouds, saving me time and money while pushing my AI work forward.

Eckart Burgwedel

CEO @Uberchord

With dstack, running LLMs on a cloud GPU is as easy as running a local Docker container. It combines the ease of Docker with the auto-scaling capabilities of K8S.

Peter Hill

Co-Founder @CUDO Compute

dstack simplifies infrastructure provisioning and AI development. If your team is on the lookout for an AI platform, I wholeheartedly recommend dstack.




Get started in minutes

Install dstack on your laptop via uv and start it using the CLI, or deploy it anywhere with the dstackai/dstack Docker image.

Set up backends or SSH fleets, then add your team.

Installation Quickstart

FAQ

How does dstack compare to Kubernetes?

Kubernetes is a widely used container orchestrator designed for general-purpose deployments. To efficiently support GPU workloads, Kubernetes typically requires custom operators, and it may not offer the most intuitive interface for ML engineers.

dstack takes a different approach, focusing on container orchestration specifically for AI workloads, with the goal of making life easier for ML engineers.

Designed to be lightweight, dstack provides a simpler, more intuitive interface for development, training, and inference. It also enables more flexible and cost-effective provisioning and management of clusters.

For optimal flexibility, dstack and Kubernetes can complement each other: dstack can handle development, while Kubernetes manages production deployments.

How does dstack differ from Slurm?

Slurm excels at job scheduling across pre-configured clusters.

dstack goes beyond scheduling, providing a full suite of features tailored to ML teams, including cluster management, dynamic compute provisioning, development environments, and advanced monitoring. This makes dstack a more comprehensive solution for AI workloads, whether in the cloud or on-prem.

When should I use dstack?

dstack is designed for ML teams aiming to speed up development while reducing GPU costs across top cloud providers or on-prem clusters.

Seamlessly integrated with Git, dstack works with any open-source or proprietary frameworks, making it developer-friendly and vendor-agnostic for training and deploying AI models.

For ML teams seeking a more streamlined, AI-native development platform, dstack provides an alternative to Kubernetes and Slurm, removing the need for MLOps or custom solutions.




dstack Enterprise

Looking for a self-hosted dstack with SSO, governance controls, and enterprise support?

dstack Sky

Don't want to host the dstack server or want to get the cheapest GPUs from the marketplace?

Already have an account? Sign in