Google Cloud Inference APIGoogle
|
||||||
Related Products
|
||||||
About
Time-series analysis is essential for the day-to-day operation of many companies. Most popular use cases include analyzing foot traffic and conversion for retailers, detecting data anomalies, identifying correlations in real-time over sensor data, or generating high-quality recommendations. With Cloud Inference API Alpha, you can gather insights in real-time from your typed time-series datasets. Get everything you need to understand your API queries results, such as groups of events that were examined, the number of groups of events, and the background probability of each returned event. Stream data in real-time, making it possible to compute correlations for real-time events. Rely on Google Cloud’s end-to-end infrastructure and defense-in-depth approach to security that’s been innovated on for over 15 years through consumer apps. At its core, Cloud Inference API is fully integrated with other Google Cloud Storage services.
|
About
NVIDIA Triton™ inference server delivers fast and scalable AI in production. Open-source inference serving software, Triton inference server streamlines AI inference by enabling teams deploy trained AI models from any framework (TensorFlow, NVIDIA TensorRT®, PyTorch, ONNX, XGBoost, Python, custom and more on any GPU- or CPU-based infrastructure (cloud, data center, or edge). Triton runs models concurrently on GPUs to maximize throughput and utilization, supports x86 and ARM CPU-based inferencing, and offers features like dynamic batching, model analyzer, model ensemble, and audio streaming. Triton helps developers deliver high-performance inference aTriton integrates with Kubernetes for orchestration and scaling, exports Prometheus metrics for monitoring, supports live model updates, and can be used in all major public cloud machine learning (ML) and managed Kubernetes platforms. Triton helps standardize model deployment in production.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
Anyone seeking a tool to run large-scale correlations over typed time-series datasets
|
Audience
Developers and companies searching for an inference server solution to improve AI production
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
No information available.
Free Version
Free Trial
|
Pricing
Free
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationGoogle
United States
cloud.google.com/inference/
|
Company InformationNVIDIA
United States
developer.nvidia.com/nvidia-triton-inference-server
|
|||||
Alternatives |
Alternatives |
|||||
|
|
||||||
|
|
||||||
|
|
|
|||||
|
|
||||||
Categories |
Categories |
|||||
Integrations
Alibaba CloudAP
Amazon EKS
Amazon SageMaker
Azure Kubernetes Service (AKS)
Azure Machine Learning
FauxPilot
Google Cloud Platform
Google Cloud Storage
Google Kubernetes Engine (GKE)
HPE Ezmeral
|
Integrations
Alibaba CloudAP
Amazon EKS
Amazon SageMaker
Azure Kubernetes Service (AKS)
Azure Machine Learning
FauxPilot
Google Cloud Platform
Google Cloud Storage
Google Kubernetes Engine (GKE)
HPE Ezmeral
|
|||||
|
|
|