Supporting GPU provisioning and orchestration on Nebius¶
As demand for GPU compute continues to scale, open-source tools tailored for AI workloads are becoming critical to
developer velocity and efficiency.
dstack
is an open-source orchestrator purpose-built for AI infrastructure—offering a lightweight, container-native
alternative to Kubernetes and Slurm.
Today, we’re announcing native integration with Nebius , offering a streamlined developer experience for teams using GPUs for AI workloads.
About Nebius¶
Nebius provides cloud GPUs, offering high-performance clusters at competitive prices. This pricing is achieved through custom-designed hardware, partnerships with Original Design Manufacturers (ODMs), and infrastructure team expertise.
Nebius offers various NVIDIA GPUs, including the L40S, H100, H200, GB200, NVL72, and B200 models, available on-demand and through reserved instances. Their data centers are located across Europe, with planned expansions into the US.
Why dstack¶
Kubernetes offers flexibility, but its complexity is often unnecessary—especially for use cases like interactive development or multi-stage training. Slurm is excellent for batch scheduling but lacks native support for dev environments, real-time inference, and multi-user orchestration.
dstack
fills the gap: a developer-friendly platform with native GPU support across dev environments, tasks, and
long-running services—without the operational overhead.
Getting started¶
To use dstack
with Nebius, configure your nebius
backend:
- Log in to your Nebius AI Cloud account.
- Navigate to
Access
, and selectService Accounts
. - Create a new service account, assign it to the
editors
group, and upload an authorized key.
Then, configure the backend via ~/.dstack/server/config.yml
:
projects:
- name: main
backends:
- type: nebius
creds:
type: service_account
service_account_id: serviceaccount-e002dwnbz81sbvg2bs
public_key_id: publickey-e00fciu5rkoteyzo69
private_key_file: ~/path/to/key.pem
Now, proceed with installing and starting the dstack
server:
$ pip install "dstack[nebius]"
$ dstack server
For more details, refer to Installation.
Use the dstack
CLI to
manage dev environments, tasks,
and services.
$ dstack apply -f .dstack.yml
# BACKEND REGION RESOURCES SPOT PRICE
1 nebius eu-north1 8xCPU, 32GB, 1xL40S (48GB) no $1.5484
2 nebius eu-north1 16xCPU, 200GB, 1xH100 (80GB) no $2.95
3 nebius eu-north1 16xCPU, 200GB, 1xH200 (141GB) no $3.5
...
Shown 3 of 7 offers, $28 max
Override the run? [y/n]:
The new nebius
backend supports CPU and GPU instances, fleets,
distributed tasks, and more.
Support for network volumes and accelerated cluster interconnects is coming soon.
What's next?
- Check Quickstart
- Sign up with Nebius AI Cloud
- Read about dev environments, tasks, services, and fleets
- Join Discord