NVIDIA DGX Cloud LeptonNVIDIA
|
||||||
Related Products
|
||||||
About
NVIDIA DGX Cloud Lepton is an AI platform that connects developers to a global network of GPU compute across multiple cloud providers through a single platform. It offers a unified experience to discover and utilize GPU resources, along with integrated AI services to streamline the deployment lifecycle across multiple clouds. Developers can start building with instant access to NVIDIA’s accelerated APIs, including serverless endpoints, prebuilt NVIDIA Blueprints, and GPU-backed compute. When it’s time to scale, DGX Cloud Lepton powers seamless customization and deployment across a global network of GPU cloud providers. It enables frictionless deployment across any GPU cloud, allowing AI applications to be deployed across multi-cloud and hybrid environments with minimal operational burden, leveraging integrated services for inference, testing, and training workloads.
|
About
NVIDIA DGX Cloud Serverless Inference is a high-performance, serverless AI inference solution that accelerates AI innovation with auto-scaling, cost-efficient GPU utilization, multi-cloud flexibility, and seamless scalability. With NVIDIA DGX Cloud Serverless Inference, you can scale down to zero instances during periods of inactivity to optimize resource utilization and reduce costs. There's no extra cost for cold-boot start times, and the system is optimized to minimize them. NVIDIA DGX Cloud Serverless Inference is powered by NVIDIA Cloud Functions (NVCF), which offers robust observability features. It allows you to integrate your preferred monitoring tools, such as Splunk, for comprehensive insights into your AI workloads. NVCF offers flexible deployment options for NIM microservices while allowing you to bring your own containers, models, and Helm charts.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
AI developers and enterprises seeking a solution to access global GPU resources for developing, training, and deploying AI models
|
Audience
Enterprises requiring a solution for deploying AI inference workloads across multi-cloud environments without the complexity of managing underlying infrastructure
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
No information available.
Free Version
Free Trial
|
Pricing
No information available.
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationNVIDIA
Founded: 1993
United States
www.nvidia.com/en-us/data-center/dgx-cloud-lepton/
|
Company InformationNVIDIA
Founded: 1993
United States
developer.nvidia.com/dgx-cloud/serverless-inference
|
|||||
Alternatives |
Alternatives |
|||||
|
|
||||||
|
|
||||||
Categories |
Categories |
|||||
Integrations
CoreWeave
Nebius
Yotta
Amazon Web Services (AWS)
Crusoe
Fluidstack
Foundry
Google Cloud Platform
HYDRA Software
Helm
|
Integrations
CoreWeave
Nebius
Yotta
Amazon Web Services (AWS)
Crusoe
Fluidstack
Foundry
Google Cloud Platform
HYDRA Software
Helm
|
|||||
|
|
|