Beyond Kubernetes: 2024 recap and what's ahead for AI infra¶
At dstack, we aim to simplify AI model development, training, and deployment of AI models by offering an
alternative to the complex Kubernetes ecosystem. Our goal is to enable seamless AI infrastructure management across any
cloud or hardware vendor.
As 2024 comes to a close, we reflect on the milestones we've achieved and look ahead to the next steps.
While dstack integrates with leading cloud GPU providers, we aim to expand partnerships with more providers
sharing our vision of simplifying AI infrastructure orchestration with a lightweight, efficient alternative to Kubernetes.
We’d also like to thank Oracle
for their collaboration, ensuring seamless integration between dstack and OCI.
Special thanks to Lambda and
Hot Aisle for providing NVIDIA and AMD hardware, enabling us conducting
benchmarks, which
are essential to advancing open-source inference and training stacks for all accelerator chips.
Thanks to your support, the project has
reached 1.6K stars on GitHub ,
reflecting the growing interest and trust in its mission.
Your issues, pull requests, as well as feedback on Discord , play a
critical role in the project's development.
Unlike Kubernetes, where node groups are typically managed through auto-scaling policies, dstack offers a more
streamlined approach. With dstack, you simply define a fleet YAML file and run
dstack apply. This command automatically provisions clusters across any cloud provider.
For quick deployments, you can skip defining a fleet altogether. When you run a dev environment, task, or service,
dstack creates a fleet automatically.
Managing on-prem resources with dstack's fleets is equally straightforward. If you have SSH access to a group of hosts, simply
list them in a YAML configuration file and run dstack apply.
type:fleet# The name is optional, if not specified, generated randomlyname:my-fleet# Ensure instances are inter-connectedplacement:cluster# The user, private SSH key, and hostnames of the on-prem serversssh_config:user:ubuntuidentity_file:~/.ssh/id_rsahosts:-3.255.177.51-3.255.177.52examples/misc/fleets/distrib-ssh.dstack.yml
This turns your on-prem cluster into a dstack fleet, ready to run dev environments, tasks, and services.
At dstack, when running a job on an instance, it uses all available GPUs on that instance. In Q1 2025, we will
introduce GPU blocks ,
allowing the allocation of instance GPUs into discrete blocks that can be reused by concurrent jobs.
This will enable more cost-efficient utilization of expensive instances.
NVIDIA remains the top accelerator supported by dstack. Recently, we introduced a NIM example
for model deployment, and we continue to enhance support for the rest of NVIDIA's ecosystem.
This year, we’re particularly proud of our newly added integration with AMD.
dstack works seamlessly with any on-prem AMD clusters. For example, you can rent such servers through our partner
Hot Aisle .
Among cloud providers, AMD is supported only through RunPod. In Q1 2025, we plan to extend it to
Nscale ,
Hot Aisle , and potentially other providers open to collaboration.
If you're interested in simplifying AI infrastructure, both in the cloud and on-prem, consider getting involved as a
dstack user, open-source contributor, or ambassador.
Finally, if you're a cloud, hardware, or software vendor, consider contributing to dstack and helping us drive it as
an open standard together.