Related Products
|
||||||
About
NVIDIA DGX Cloud Serverless Inference is a high-performance, serverless AI inference solution that accelerates AI innovation with auto-scaling, cost-efficient GPU utilization, multi-cloud flexibility, and seamless scalability. With NVIDIA DGX Cloud Serverless Inference, you can scale down to zero instances during periods of inactivity to optimize resource utilization and reduce costs. There's no extra cost for cold-boot start times, and the system is optimized to minimize them. NVIDIA DGX Cloud Serverless Inference is powered by NVIDIA Cloud Functions (NVCF), which offers robust observability features. It allows you to integrate your preferred monitoring tools, such as Splunk, for comprehensive insights into your AI workloads. NVCF offers flexible deployment options for NIM microservices while allowing you to bring your own containers, models, and Helm charts.
|
About
Oracle Cloud Infrastructure provides fast, flexible, and affordable compute capacity to fit any workload need from performant bare metal servers and VMs to lightweight containers. OCI Compute provides uniquely flexible VM and bare metal instances for optimal price-performance. Select exactly the number of cores and the memory your applications need. Delivering high performance for enterprise workloads. Simplify application development with serverless computing. Your choice of technologies includes Kubernetes and containers. NVIDIA GPUs for machine learning, scientific visualization, and other graphics processing. Capabilities such as RDMA, high-performance storage, and network traffic isolation. Oracle Cloud Infrastructure consistently delivers better price performance than other cloud providers. Virtual machine-based (VM) shapes offer customizable core and memory combinations. Customers can optimize costs by choosing a specific number of cores.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
Enterprises requiring a solution for deploying AI inference workloads across multi-cloud environments without the complexity of managing underlying infrastructure
|
Audience
Professionals seeking a tool to solve their workload needs from performant bare metal servers and VMs to lightweight containers
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
No information available.
Free Version
Free Trial
|
Pricing
$0.007 per hour
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationNVIDIA
Founded: 1993
United States
developer.nvidia.com/dgx-cloud/serverless-inference
|
Company InformationOracle
United States
www.oracle.com/cloud/compute/
|
|||||
Alternatives |
Alternatives |
|||||
|
|
||||||
Categories |
Categories |
|||||
Integrations
Oracle Cloud Infrastructure
Amazon Web Services (AWS)
CentOS
CoreWeave
Google Cloud Platform
Helm
Kubernetes
Llama
Microsoft Azure
NVIDIA AI Foundations
|
Integrations
Oracle Cloud Infrastructure
Amazon Web Services (AWS)
CentOS
CoreWeave
Google Cloud Platform
Helm
Kubernetes
Llama
Microsoft Azure
NVIDIA AI Foundations
|
|||||
|
|
|