The orchestration layer for modern ML teams

dstack gives your team a single control plane to run development, training, and inference jobs on GPU—whether on hyperscalers, neoclouds, or your on-prem hardware. Avoid vendor lock-in and minimize GPU spend.

The entire setup takes minutes.

One control plane for all your GPUs

Instead of wrestling with complex Helm charts and Kubernetes operators, dstack provides a simple, declarative way to manage clusters, containerized dev environments, training, and inference.

This container-native interface makes your team more productive and your GPU usage more efficient—leading to lower costs and faster iteration.

Run on any cloud. Or all of them.

dstack natively provisions GPU instances and clusters across your preferred cloud providers—maximizing efficiency and minimizing overhead.

Connect GPU clouds directly to dstack natively, or run dstack on top of Kubernetes if needed.

Backends

Bring your own compute

Have bare-metal servers or on-prem GPU clusters? dstack makes it easy to integrate them and manage everything alongside your cloud resources.

With SSH fleets, you can connect any existing cluster in minutes. Once added to dstack, it's a first-class resource — available for dev environments, tasks, and services.

SSH fleets

Code locally, run remotely

Before training or deployment, ML engineers explore and debug their code.

dstack's dev environments make it easy to connect your desktop IDE to powerful cloud or on-prem GPUs—streamlining the entire development loop.

Dev environments

Run complex training with simple config

Move from single-instance experiments to multi-node distributed training without friction. dstack lets you define complex jobs with a simple configuration, handling the scheduling and orchestration for you.

This allows your team to focus on research while ensuring that expensive cluster resources are utilized efficiently.

Tasks

Deploy scalable, OpenAI-compatible endpoints

With dstack, you can easily deploy any model as a secure, auto-scaling OpenAI-compatible endpoint, all while using your custom code, Docker image, and serving framework.

Services

Loved by world-class ML teams

Wah Loon Keng

Sr. AI Engineer @Electronic Arts

With dstack, AI researchers at EA can spin up and scale experiments without touching infrastructure. It supports everything from quick prototyping to multi-node training on any cloud.

Aleksandr Movchan

ML Engineer @Mobius Labs

Thanks to dstack, my team can quickly tap into affordable GPUs and streamline our workflows from testing and development to full-scale application deployment.

Alvaro Bartolome

ML Engineer @Argilla

With dstack it's incredibly easy to define a configuration within a repository and run it without worrying about GPU availability. It lets you focus on data and your research.

Park Chansung

ML Researcher @ETRI

Thanks to dstack, I can effortlessly access the top GPU options across different clouds, saving me time and money while pushing my AI work forward.

Eckart Burgwedel

CEO @Uberchord

With dstack, running LLMs on a cloud GPU is as easy as running a local Docker container. It combines the ease of Docker with the auto-scaling capabilities of K8S.

Peter Hill

Co-Founder @CUDO Compute

dstack simplifies infrastructure provisioning and AI development. If your team is on the lookout for an AI platform, I wholeheartedly recommend dstack.

FAQ

How does dstack compare to Kubernetes?

Kubernetes is a widely used container orchestrator designed for general-purpose deployments. To efficiently support GPU workloads, Kubernetes typically requires custom operators, and it may not offer the most intuitive interface for ML engineers.

dstack takes a different approach, focusing on container orchestration specifically for AI workloads, with the goal of making life easier for ML engineers.

Designed to be lightweight, dstack provides a simpler, more intuitive interface for development, training, and inference. It also enables more flexible and cost-effective provisioning and management of clusters.

For optimal flexibility, dstack and Kubernetes can complement each other: dstack can handle development, while Kubernetes manages production deployments.

How does dstack differ from Slurm?

Slurm excels at job scheduling across pre-configured clusters.

dstack goes beyond scheduling, providing a full suite of features tailored to ML teams, including cluster management, dynamic compute provisioning, development environments, and advanced monitoring. This makes dstack a more comprehensive solution for AI workloads, whether in the cloud or on-prem.

When should I use dstack?

dstack is designed for ML teams aiming to speed up development while reducing GPU costs across top cloud providers or on-prem clusters.

Seamlessly integrated with Git, dstack works with any open-source or proprietary frameworks, making it developer-friendly and vendor-agnostic for training and deploying AI models.

For ML teams seeking a more streamlined, AI-native development platform, dstack provides an alternative to Kubernetes and Slurm, removing the need for MLOps or custom solutions.

Have questions, or need help?
Discord Talk to us




Get started in minutes

Install dstack on your laptop via uv and start it using the CLI, or deploy it anywhere with the dstackai/dstack Docker image.

Set up backends or SSH fleets, then add your team.

Installation Quickstart

dstack Sky

Don't want to host the dstack server or want to get the cheapest GPUs from the marketplace?

Already have an account? Sign in

dstack Enterprise

Looking for a self-hosted dstack with SSO, governance controls, and enterprise support?

Talk to our experts and learn.