NVIDIA and Google Cloud
NVIDIA and Google Cloud deliver accelerator-optimized solutions that address your most demanding workloads, including machine learning, high performance computing, data analytics, graphics, and gaming workloads.
The power of NVIDIA-accelerated computing at scale on Google Cloud
Increased performance for diverse workloads
With the latest NVIDIA GPUs on Google Cloud, you can easily provision Compute Engine instances with NVIDIA A100, P100, P4, T4 or V100 to accelerate your most demanding workloads.
Reduce costs with per-second billing
Google Cloud's per-second pricing means you pay only for what you need, with up to a 30% monthly discount applied automatically. Save on up-front costs while enjoying the same uptime and scalable performance.
Optimize workloads with custom machine configurations
Optimize your workloads by precisely configuring an instance with the exact ratio of processors, memory, and NVIDIA GPUs you need instead of modifying your workload to fit within limited system configurations.
NVIDIA technologies on Google Cloud
NVIDIA A100® Tensor Core GPU
The accelerator-optimized A2 VMs are based on the NVIDIA Ampere A100 Tensor Core GPU. Each A100 GPU offers up to 20x the compute performance of the previous generation. These VMs are designed to deliver acceleration at every scale for AI, data analytics, and high performance computing to tackle the toughest computing challenges.
NVIDIA T4® Tensor Core GPU
The NVIDIA® T4 GPU accelerates diverse cloud workloads, including high-performance computing, deep learning training and inference, machine learning, data analytics, and graphics. NVIDIA® T4 GPUs are generally available on Compute Engine.
Hybrid cloud with NVIDIA and Google Cloud’s Anthos
Google Cloud's Anthos was built to enable customers to easily run applications both in the cloud and on-premises. Working closely with NVIDIA, we've built a solution that uses the NVIDIA GPU Operator to deploy the components required to enable GPUs in Kubernetes. The solution works with many popular NVIDIA GPUs, including the A100 and T4.
Autoscale with Google Kubernetes Engine
Using Google Kubernetes Engine (GKE) you can seamlessly create clusters with NVIDIA GPUs on demand, load balance, and minimize operational costs by automatically scaling GPU resources up or down. With support for multi-instance GPUs (MIG) in NVIDIA A100 GPUs, GKE can now provision the right-size GPU acceleration with finer granularity for multiuser, multimodel AI inference workloads.
NVIDIA CloudXR™ with RTX Virtual Workstations
NVIDIA CloudXR, a groundbreaking innovation built on NVIDIA RTX™ technology, makes high-quality XR accessible through Google Cloud Marketplace with NVIDIA RTX Virtual Workstation as a virtual machine image (VMI). Users can easily set up, scale, and consume high-quality immersive experience and stream XR workflows from the cloud.
Customers benefiting from NVIDIA GPU acceleration in Google Cloud
Google Cloud offerings with NVIDIA technologies
Technical resources for deploying NVIDIA technologies on Google Cloud
GPUs on Compute Engine
Compute Engine provides GPUs that you can add to your virtual machine instances. Learn what you can do with GPUs and what types of GPU hardware are available.
Using GPUs for training models in the cloud
Accelerate the training process for many deep-learning models, like image classification, video analysis, and natural language processing.
GPUs on Google Kubernetes Engine
Learn how to use GPU hardware accelerators in your Google Kubernetes Engine clusters’ nodes.
Attaching GPUs to Dataproc clusters
Attach GPUs to the master and worker Compute Engine nodes in a Dataproc cluster to accelerate specific workloads, such as machine learning and data processing.