About
NVIDIA DGX Cloud Lepton is a GPU compute marketplace and developer platform that aggregates accelerated computing resources from a global network of cloud providers into a single, unified access point. Designed for AI developers, researchers, and enterprises, it removes the complexity of managing multi-cloud GPU infrastructure by offering seamless access to NVIDIA-accelerated hardware at scale. The platform supports a wide range of AI workloads including large language model training and inference, fine-tuning, data processing, and high-performance computing tasks. Built on NVIDIA's deep expertise in AI infrastructure — including Blackwell and Hopper GPU architectures, NVLink, and Tensor Core technologies — DGX Cloud Lepton gives developers programmatic access to the compute they need without being locked into a single cloud provider. This makes it especially valuable for startups and enterprises that need flexible, scalable GPU access as their AI workloads grow. The platform is integrated into NVIDIA's broader DGX Cloud ecosystem, which includes tools for AI factory deployment, MLOps orchestration via NVIDIA Run:ai, and enterprise AI software through NVIDIA AI Enterprise. Developers can access Lepton through a web interface or API, making it easy to integrate into existing ML pipelines and workflows.
Key Features
- Multi-Cloud GPU Marketplace: Aggregates GPU compute from multiple cloud providers worldwide into a single platform, giving developers flexibility and availability without vendor lock-in.
- NVIDIA-Accelerated Hardware: Access to NVIDIA's latest GPU architectures including Blackwell and Hopper, with support for NVLink, Tensor Cores, and multi-instance GPU configurations.
- Developer-First API Access: Provides programmatic API access to GPU compute resources, enabling easy integration into existing ML pipelines, training frameworks, and deployment workflows.
- Scalable AI Infrastructure: Supports a wide range of workloads from LLM training and inference to HPC simulations, with the ability to scale compute up or down on demand.
- DGX Ecosystem Integration: Seamlessly integrates with NVIDIA's broader AI platform including AI Enterprise software, Run:ai orchestration, and Base Command Manager for end-to-end AI factory operations.
Use Cases
- AI startups that need on-demand GPU compute for LLM training and fine-tuning without committing to a single cloud provider's capacity constraints.
- Enterprise ML teams running large-scale model inference workloads who need scalable, high-throughput GPU infrastructure with enterprise SLAs.
- Researchers conducting HPC simulations or deep learning experiments that require burst GPU capacity beyond what a single cloud can reliably provide.
- MLOps teams building production AI pipelines who need programmatic API access to GPU resources integrated with orchestration and monitoring tools.
- Organizations evaluating multi-cloud AI infrastructure strategies who want to compare GPU availability and pricing across providers from a single platform.
Pros
- No Vendor Lock-In: Access GPU compute from multiple global cloud providers through one platform, making it easy to switch providers or distribute workloads for cost and availability optimization.
- Latest NVIDIA Hardware: Direct access to NVIDIA's most advanced GPU architectures ensures developers can run cutting-edge AI workloads with maximum performance.
- Enterprise-Grade Ecosystem: Built on NVIDIA's trusted infrastructure stack, with support for MLOps, virtualization, confidential computing, and enterprise AI software.
Cons
- Primarily Paid: As a cloud compute platform, all GPU access is metered and paid — there is no meaningful free tier for sustained AI development workloads.
- Complexity for Beginners: The platform is designed for professional developers and enterprise teams; those new to GPU computing or cloud infrastructure may face a steep learning curve.
- NVIDIA Ecosystem Dependency: While multi-cloud, the platform is centered on NVIDIA GPU hardware, which may not suit teams that require AMD or other accelerator types.
Frequently Asked Questions
NVIDIA DGX Cloud Lepton is a platform that connects AI developers to a global network of GPU compute resources across multiple cloud providers, all accessible through a single unified interface.
Lepton is designed for AI developers, ML engineers, researchers, and enterprise teams that need scalable GPU compute to train, fine-tune, or deploy large AI models and run HPC workloads.
Lepton provides access to NVIDIA's latest GPU architectures including Blackwell (GB200, GB300) and Hopper (H100), along with NVLink-connected multi-GPU configurations for large-scale AI training and inference.
Unlike using a single cloud provider, Lepton aggregates GPU supply from multiple global clouds, reducing availability bottlenecks, enabling cost comparisons, and eliminating vendor lock-in — all through one API and interface.
Yes. Lepton offers API access that can be integrated into existing ML pipelines, training frameworks, and orchestration tools. It also integrates with NVIDIA's Run:ai for GPU workload orchestration and NVIDIA AI Enterprise for software tooling.
