Accelerate innovation on Google Cloud and NVIDIA
NVIDIA and Google Cloud provide accelerator-optimized solutions that support demanding workloads, including generative AI, high-performance computing, data analytics, graphics, and gaming workloads.
Recapping on NVIDIA GTC 2026 San Jose. See the highlights:
“We are moving from training AI to producing intelligence. These data centers are no longer just storing information. They are factories generating tokens, generating intelligence... Our expanded collaboration with Google Cloud will help developers accelerate their work with infrastructure that supercharges energy efficiency and reduces costs.”
Jensen Huang, GTC 2026 keynote
High Performing GPUs on Google Cloud
Speed up machine learning, scientific computing, and generative AI with high-performance GPUs on Google Cloud.
Key benefits:
Key features
NVIDIA technologies on Google Cloud
Google Kubernetes Engine (GKE)
Use GKE scalability, NVIDIA Multi-Instance GPU (MIG) support, and GPU time-sharing for efficient generative AI training, inference, and other compute-intensive workloads. Optimize resource utilization and minimize operational costs.
Vertex AI
Combine NVIDIA accelerated computing with Vertex AI, a unified MLOps platform. Utilize NVIDIA GPUs and AI software (such as, Triton™ Inference Server) within Vertex AI Training, Prediction, Pipelines, and Notebooks to accelerate generative AI development and deployment without infrastructure complexities.
Cloud Run
Deploy generative AI faster with NVIDIA NIM on Cloud Run, a fully managed serverless platform. Cloud Run GPU support speeds up NIM to optimize performance and accelerate gen AI model deployment in a serverless environment.
Dynamic Workload Scheduler
Access NVIDIA GPU capacity on Google Cloud for short-duration AI workloads (training, fine-tuning, experimentation). Flexible scheduling and atomic provisioning enhance resource utilization and optimize costs across services like GKE, Vertex AI, and Batch.
Google Distributed Cloud
The NVIDIA Blackwell platform on Google Distributed Cloud enables secure, on-premises deployment of advanced agentic AI (including Google Gemini models). This offers improved AI performance and scalability for sensitive, regulated workloads, ensuring data privacy, sovereignty, and compliance.
Technical resources for deploying NVIDIA technologies on Google Cloud
Google Cloud basics
Tutorials
Recapping on NVIDIA GTC 2026 San Jose. See the highlights:
“We are moving from training AI to producing intelligence. These data centers are no longer just storing information. They are factories generating tokens, generating intelligence... Our expanded collaboration with Google Cloud will help developers accelerate their work with infrastructure that supercharges energy efficiency and reduces costs.”
Jensen Huang, GTC 2026 keynote
High Performing GPUs on Google Cloud
Speed up machine learning, scientific computing, and generative AI with high-performance GPUs on Google Cloud.
Key benefits:
Key features
NVIDIA technologies on Google Cloud
Google Kubernetes Engine (GKE)
Use GKE scalability, NVIDIA Multi-Instance GPU (MIG) support, and GPU time-sharing for efficient generative AI training, inference, and other compute-intensive workloads. Optimize resource utilization and minimize operational costs.
Vertex AI
Combine NVIDIA accelerated computing with Vertex AI, a unified MLOps platform. Utilize NVIDIA GPUs and AI software (such as, Triton™ Inference Server) within Vertex AI Training, Prediction, Pipelines, and Notebooks to accelerate generative AI development and deployment without infrastructure complexities.
Cloud Run
Deploy generative AI faster with NVIDIA NIM on Cloud Run, a fully managed serverless platform. Cloud Run GPU support speeds up NIM to optimize performance and accelerate gen AI model deployment in a serverless environment.
Dynamic Workload Scheduler
Access NVIDIA GPU capacity on Google Cloud for short-duration AI workloads (training, fine-tuning, experimentation). Flexible scheduling and atomic provisioning enhance resource utilization and optimize costs across services like GKE, Vertex AI, and Batch.
Google Distributed Cloud
The NVIDIA Blackwell platform on Google Distributed Cloud enables secure, on-premises deployment of advanced agentic AI (including Google Gemini models). This offers improved AI performance and scalability for sensitive, regulated workloads, ensuring data privacy, sovereignty, and compliance.
Technical resources for deploying NVIDIA technologies on Google Cloud
Google Cloud basics
Tutorials





Compared with another inference platform, running on GKE with NVIDIA NIM and GPUs delivered 6.1x acceleration in average answer/response generation speed for the Amazfit AI agent.
Jia Li Co-Founder, Chief AI Officer, LiveX AI