Skip to content

dstack is an open-source alternative to Kubernetes, designed to simplify development, training, and deployment of AI across any cloud or on-prem. It supports NVIDIA, AMD, and TPU.

License

Notifications You must be signed in to change notification settings

Innovatorcloudy/dstack

 
 

Repository files navigation

dstack is a streamlined alternative to Kubernetes, specifically designed for AI. It simplifies container orchestration for AI workloads both in the cloud and on-prem, speeding up the development, training, and deployment of AI models.

dstack is easy to use with any cloud provider as well as on-prem servers.

Accelerators

dstack supports NVIDIA GPU, AMD GPU, and Google Cloud TPU out of the box.

Major news ✨

Installation

Before using dstack through CLI or API, set up a dstack server. If you already have a running dstack server, you only need to set up the CLI.

(Optional) Configure backends

To use dstack with your own cloud accounts, create the ~/.dstack/server/config.yml file and configure backends. Alternatively, you can configure backends via the control plane UI after you start the server.

You can skip backends configuration if you intend to run containers only on your on-prem servers. Use SSH fleets for that.

Start the server

Once the backends are configured, proceed to start the server:

$ pip install "dstack[all]" -U
$ dstack server

Applying ~/.dstack/server/config.yml...

The admin token is "bbae0f28-d3dd-4820-bf61-8f4bb40815da"
The server is running at http://127.0.0.1:3000/

For more details on server configuration options, see the server deployment guide.

Set up the CLI

To point the CLI to the dstack server, configure it with the server address, user token, and project name:

$ pip install dstack
$ dstack config --url http://127.0.0.1:3000 \
    --project main \
    --token bbae0f28-d3dd-4820-bf61-8f4bb40815da
    
Configuration is updated at ~/.dstack/config.yml

How does it work?

1. Define configurations

dstack supports the following configurations:

  • Dev environments — for interactive development using a desktop IDE
  • Tasks — for scheduling jobs (incl. distributed jobs) or running web apps
  • Services — for deployment of models and web apps (with auto-scaling and authorization)
  • Fleets — for managing cloud and on-prem clusters
  • Volumes — for managing persisted volumes
  • Gateways — for configuring the ingress traffic and public endpoints

Configuration can be defined as YAML files within your repo.

2. Apply configurations

Apply the configuration either via the dstack apply CLI command or through a programmatic API.

dstack automatically manages provisioning, job queuing, auto-scaling, networking, volumes, run failures, out-of-capacity errors, port-forwarding, and more — across clouds and on-prem clusters.

More information

For additional information and examples, see the following links:

Contributing

You're very welcome to contribute to dstack. Learn more about how to contribute to the project at CONTRIBUTING.md.

License

Mozilla Public License 2.0

About

dstack is an open-source alternative to Kubernetes, designed to simplify development, training, and deployment of AI across any cloud or on-prem. It supports NVIDIA, AMD, and TPU.

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 74.8%
  • TypeScript 17.8%
  • Go 5.5%
  • Shell 0.6%
  • SCSS 0.5%
  • JavaScript 0.5%
  • Other 0.3%