The Applied ML Summit is coming up on June 9! Join us to learn how you can speed up experimentation and get into production quickly.

Jump to

NVIDIA and Google Cloud

NVIDIA and Google Cloud deliver accelerator-optimized solutions that address your most demanding workloads, including machine learning, high performance computing, data analytics, graphics, and gaming workloads.  

Nvidia and Google Cloud logos


The power of NVIDIA-accelerated computing at scale on Google Cloud

Increased performance for diverse workloads

With the latest NVIDIA GPUs on Google Cloud, you can easily provision Compute Engine instances with NVIDIA A100, P100, P4, T4 or V100 to accelerate your most demanding workloads.

Reduce costs with per-second billing

Google Cloud's per-second pricing means you pay only for what you need, with up to a 30% monthly discount applied automatically. Save on up-front costs while enjoying the same uptime and scalable performance.

Optimize workloads with custom machine configurations

Optimize your workloads by precisely configuring an instance with the exact ratio of processors, memory, and NVIDIA GPUs you need instead of modifying your workload to fit within limited system configurations. 

Key features

NVIDIA technologies on Google Cloud

NVIDIA A100® Tensor Core GPU

The accelerator-optimized A2 VMs are based on the NVIDIA Ampere A100 Tensor Core GPU. Each A100 GPU offers up to 20x the compute performance of the previous generation. These VMs are designed to deliver acceleration at every scale for AI, data analytics, and high performance computing to tackle the toughest computing challenges.

NVIDIA T4® Tensor Core GPU

The NVIDIA® T4 GPU accelerates diverse cloud workloads, including high-performance computing, deep learning training and inference, machine learning, data analytics, and graphics. NVIDIA® T4 GPUs are generally available on Compute Engine.

Hybrid cloud with NVIDIA and Google Cloud’s Anthos

Google Cloud's Anthos was built to enable customers to easily run applications both in the cloud and on-premises. Working closely with NVIDIA, we've built a solution that uses the NVIDIA GPU Operator to deploy the components required to enable GPUs in Kubernetes. The solution works with many popular NVIDIA GPUs, including the A100 and T4.

Autoscale with Google Kubernetes Engine

Using Google Kubernetes Engine (GKE) you can seamlessly create clusters with NVIDIA GPUs on demand, load balance, and minimize operational costs by automatically scaling GPU resources up or down. With support for multi-instance GPUs (MIG) in NVIDIA A100 GPUs, GKE can now provision the right-size GPU acceleration with finer granularity for multiuser, multimodel AI inference workloads.

NVIDIA CloudXR™ with RTX Virtual Workstations

NVIDIA CloudXR, a groundbreaking innovation built on NVIDIA RTX™ technology, makes high-quality XR accessible through Google Cloud Marketplace with NVIDIA RTX Virtual Workstation as a virtual machine image (VMI). Users can easily set up, scale, and consume high-quality immersive experience and stream XR workflows from the cloud.

Ready to get started? Contact us


Technical resources for deploying NVIDIA technologies on Google Cloud

Google Cloud Basics
GPUs on Compute Engine

Compute Engine provides GPUs that you can add to your virtual machine instances. Learn what you can do with GPUs and what types of GPU hardware are available.

Google Cloud Basics
Using GPUs for training models in the cloud

Accelerate the training process for many deep-learning models, like image classification, video analysis, and natural language processing.

GPUs on Google Kubernetes Engine

Learn how to use GPU hardware accelerators in your Google Kubernetes Engine clusters’ nodes.

Google Cloud Basics
Attaching GPUs to Dataproc clusters

Attach GPUs to the master and worker Compute Engine nodes in a Dataproc cluster to accelerate specific workloads, such as machine learning and data processing.