
TensorFlow Model Garden
A repository of state-of-the-art model implementations for TensorFlow users.
Open-source GPU-native orchestration for AI teams.
dstack is an open-source platform designed to streamline GPU orchestration for AI and ML teams. It enables provisioning of GPUs and orchestrates containerized workloads across various environments, including cloud, Kubernetes, and bare-metal clusters. dstack focuses on increasing GPU utilization and reducing vendor lock-in. The platform supports development, distributed training, and high-throughput inference, offering a unified control plane tightly integrated with open-source frameworks. With native integration for leading GPU clouds and support for on-prem clusters via Kubernetes or SSH fleets, dstack provides flexible and efficient resource management. It also offers features like dev environments for interactive GPU access, scalable model deployment as auto-scaling endpoints, and detailed GPU utilization reporting.
Open-source GPU-native orchestration for AI teams.
Quick visual proof for dstack. Helps non-technical users understand the interface faster.
dstack is an open-source platform designed to streamline GPU orchestration for AI and ML teams.
Explore all tools that specialize in gpu orchestration. This domain focus ensures dstack delivers optimized results for this specific requirement.
Explore all tools that specialize in model training. This domain focus ensures dstack delivers optimized results for this specific requirement.
Explore all tools that specialize in model inference. This domain focus ensures dstack delivers optimized results for this specific requirement.
Open side-by-side comparison first, then move to deeper alternatives guidance.
Direct provisioning and management of GPU VMs through cloud APIs, supporting fast and efficient resource allocation.
Interactive GPU access via desktop IDEs (VS Code, Cursor) connected to cloud or on-prem GPUs for experimentation and debugging.
Deployment of models as secure, auto-scaling endpoints with OpenAI-compatible APIs, supporting disaggregated prefill/decode and cache-aware routing.
A single interface to manage GPU resources and workloads across different environments (cloud, on-prem, Kubernetes).
Directly orchestrates GPUs on bare-metal servers or VMs without Kubernetes, providing a lightweight alternative for resource management.
Connect existing Kubernetes clusters to dstack, leveraging existing infrastructure for GPU orchestration.
Allows efficient resource reuse, right-sizing, and support for spot, on-demand, and reserved capacity, radically reducing GPU costs.
Install dstack via uv or Docker.
Set up backends (Kubernetes, SSH fleets).
Configure .dstack.yml for dev environments, tasks, and services.
Use CLI commands like 'dstack apply' to deploy workloads.
All Set
Ready to go
Verified feedback from other users.
“Users praise dstack for its ease of use, seamless integration with existing infrastructure, and ability to reduce GPU costs.”
0No reviews yet. Be the first to rate this tool.

A repository of state-of-the-art model implementations for TensorFlow users.

Inference Llama 2 in one file of pure C.

The Private Cloud Infrastructure for Sovereign Generative AI.

An end-to-end open source platform for machine learning.
Supervise.ly provides an all-in-one platform for computer vision, enabling users to curate, label, train, evaluate, and deploy models for images, videos, 3D, and medical data.

The end-to-end AI cloud that simplifies building and deploying models.