Related Products
|
||||||
About
Powerful, self-serve machine learning platform where you can turn models into scalable APIs in just a few clicks. Sign up for Deep Infra account using GitHub or log in using GitHub. Choose among hundreds of the most popular ML models. Use a simple rest API to call your model. Deploy models to production faster and cheaper with our serverless GPUs than developing the infrastructure yourself. We have different pricing models depending on the model used. Some of our language models offer per-token pricing. Most other models are billed for inference execution time. With this pricing model, you only pay for what you use. There are no long-term contracts or upfront costs, and you can easily scale up and down as your business needs change. All models run on A100 GPUs, optimized for inference performance and low latency. Our system will automatically scale the model based on your needs.
|
About
NVIDIA DGX Cloud Serverless Inference is a high-performance, serverless AI inference solution that accelerates AI innovation with auto-scaling, cost-efficient GPU utilization, multi-cloud flexibility, and seamless scalability. With NVIDIA DGX Cloud Serverless Inference, you can scale down to zero instances during periods of inactivity to optimize resource utilization and reduce costs. There's no extra cost for cold-boot start times, and the system is optimized to minimize them. NVIDIA DGX Cloud Serverless Inference is powered by NVIDIA Cloud Functions (NVCF), which offers robust observability features. It allows you to integrate your preferred monitoring tools, such as Splunk, for comprehensive insights into your AI workloads. NVCF offers flexible deployment options for NIM microservices while allowing you to bring your own containers, models, and Helm charts.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
Anyone in search of a solution to run the top AI models to improve their machine learning outcomes
|
Audience
Enterprises requiring a solution for deploying AI inference workloads across multi-cloud environments without the complexity of managing underlying infrastructure
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
$0.70 per 1M input tokens
Free Version
Free Trial
|
Pricing
No information available.
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationDeep Infra
deepinfra.com
|
Company InformationNVIDIA
Founded: 1993
United States
developer.nvidia.com/dgx-cloud/serverless-inference
|
|||||
Alternatives |
Alternatives |
|||||
|
|
||||||
Categories |
Categories |
|||||
Integrations
Llama
AI SpendOps
Codestral Mamba
GitHub
Google Cloud Platform
Le Chat
Llama 3
Mathstral
Microsoft Azure
Ministral 3B
|
Integrations
Llama
AI SpendOps
Codestral Mamba
GitHub
Google Cloud Platform
Le Chat
Llama 3
Mathstral
Microsoft Azure
Ministral 3B
|
|||||
|
|
|