NVIDIA and Google Cloud Platform

NVIDIA Tesla T4 GPUs now available on Google Cloud Platform.
Watch this on-demand webinar to learn more about T4 GPUs on GCP.

Contact Us

Google Cloud Platform offers NVIDIA Tesla K80, P4, T4, P100, and V100 GPUs

Predicting our climate’s future. A new drug to treat cancer. Some of the world’s most important challenges need to be solved today, but require tremendous amounts of computing to become a reality.

Together, NVIDIA and Google Cloud are helping you achieve faster results to address these challenges — all without massive capital expenditures or complexity of managing the underlying infrastructure. The NVIDIA Tesla K80, P4, T4, P100, and V100 GPUs are all generally available on Google Cloud Platform. With NVIDIA GPUs on Google Cloud Platform, deep learning, analytics, physical simulation, video transcoding, and molecular modeling take hours instead of days. You can also leverage NVIDIA GRID virtual workstations on Google Cloud Platform to accelerate your graphics-intensive workloads from anywhere.


Benefits of using NVIDIA GPUs on Google Cloud Platform

Increased performance for complex computing
Increase the speed of your complex, compute-intensive workloads such as machine learning training and inference, medical analysis, seismic exploration, video transcoding, virtual workstations, and scientific simulations. Easily provision Google Compute Engine instances with NVIDIA Tesla K80, P4, P100, T4, or V100 to handle your most complex compute-intensive workloads.
Reduce costs with per-second billing
With Google Cloud Platform’s per-second pricing, you pay only for what you need, with up to a 30% monthly discount applied automatically. Save on up-front capital expenditure while enjoying the 24/7 uptime and scalable performance you have come to expect from NVIDIA Tesla GPUs.
Optimize workloads with custom machine configurations
Optimize your workloads by precisely configuring an instance with the ratio of processors, memory, and NVIDIA GPUs you need instead of modifying your workload to fit within limited system configurations. You can also benefit by leveraging GPU-optimized containers from the NVIDIA GPU Cloud container registry to accelerate your DL and HPC workloads on Google Cloud Platform.
Integrate seamlessly with Vertex AI
Tackle the explosion of data generated every day by transactional records, sensor logs, images, videos, and more. With NVIDIA GPU-accelerated cloud computing resources, you can generate insights from your data without the need to move it out of the cloud. NVIDIA Tesla K80, P4, P100, and V100 GPUs are tightly integrated with Vertex AI, dramatically slashing the time to train machine learning (ML) models on large datasets using TensorFlow framework and tightly integrating with Cloud Dataflow, BigQuery, Cloud Storage, and Notebooks.
Accelerate ML training time and deliver efficient ML inference
Solving today's complex challenges with ML requires training exponentially more complex deep learning models in a practical amount of time. NVIDIA Tesla T4 and V100 GPUs on Google Cloud Platform dramatically reduce training time for these models from weeks to a few hours and also provide greater efficiency when running those trained models for inference by providing an order of magnitude higher throughput with low latency for better user experiences.
Build on Google’s global infrastructure

Access some of the same hardware that Google uses to develop high-performance deep-learning products, without having to worry about the capital expenditures or IT operations of managing your own infrastructure. NVIDIA Tesla K80, P4, P100, T4, and V100 GPUs on Google Cloud Platform means the hardware is passed through directly to the virtual machine to provide bare metal performance.

Virtual workstations with NVIDIA GRID and Tesla P4, T4 and P100 GPUs enable creative and technical professionals to access demanding applications from the cloud.

NVIDIA GPUs available on Google Cloud Platform

NVIDIA Tesla K80 GPUs on Google Cloud Platform

NVIDIA Tesla K80 is generally available in the Google Cloud. It drastically lowers model-training times and HPC cost by delivering superior performance with fewer, more powerful server instances, engineered to deliver above 5–10x performance boost on real-world applications.

Over 550 industry-leading HPC applications already support NVIDIA GPUs, including all top 15 HPC applications and all deep-learning frameworks. With features like dual-GPU design and Dynamic GPU Boost, Tesla K80 is built to deliver superior performance for these applications.

Try NVIDIA K80 on Compute Engine
NVIDIA Tesla P4 and T4 GPUs on Google Cloud Platform

NVIDIA Tesla P4 and T4 GPUs are now generally available on Compute Engine.

Inference platform

NVIDIA Tesla P4 and T4 GPUs boost the efficiency of scale-out servers running deep-learning workloads and enable responsive AI-based services. The P4 and T4 are designed to slash inference latency, while providing better energy efficiency. This helps unlock AI services that were previously impossible due to latency limitations.

Virtual workstations

NVIDIA Tesla P4 and T4 help provide designers, engineers, scientists, and graphic artists virtual workstations possessing the power to take on the biggest visualization challenges with immersive, interactive, photorealistic environments. With NVIDIA GRID, subject matter experts are now untethered from their desks and can access even the most demanding professional applications and data from nearly anywhere, on virtually any device.

Video transcoding

NVIDIA Tesla P4 can transcode and infer up to 18 Full HD video streams in real time and T4 up to 38 Full HD videos streams, powered by a dedicated hardware-accelerated decode engine that works in parallel with the GPU doing inference. By integrating deep learning into the video pipeline, customers can offer smart, innovative video services to users that were previously impossible to do.

Try NVIDIA P4 on Compute Engine
NVIDIA Tesla P100 GPUs on Google Cloud Platform

NVIDIA Tesla P100 is generally available on Google Cloud Platform.

Unified supercomputing

Designed to boost throughput and save money for both HPC and ML applications. Powered by the NVIDIA Pascal architecture, each Tesla P100 delivers 4.7 and 9.3 teraflops of double-precision and single-precision performance for HPC and ML workloads.

Greater efficiency with CoWoS with HBM2

Applications often spend more time and energy waiting for data than processing it. The NVIDIA Tesla P100 tightly integrates compute and data on the same package by adding Chip on Wafer on Substrate (CoWoS) with HBM2 technology to deliver unprecedented computational efficiency. This integration provides a huge generational leap in application performance by delivering up to 3X memory bandwidth over prior-generation solutions.

Simplified parallel programming with the Page Migration Engine

Parallel programming just got a lot simpler with the Pascal architecture. The Page Migration Engine frees developers to focus more on tuning for computing performance and less on managing data movement. It also allows applications to scale beyond the physical memory size of the GPU, with support for virtual memory paging. With Unified Memory technology, developers see a single memory space for the entire instance to dramatically improve productivity.

Try NVIDIA P100 on Compute Engine
NVIDIA Tesla V100 GPU on Google Cloud Platform

NVIDIA Tesla V100 GPUs are now generally available on Compute Engine and Google Kubernetes Engine.

Today’s most demanding workloads and industries require the fastest hardware accelerators. Customers can now select as many as eight NVIDIA Tesla V100 GPUs, 96 vCPU, and 624 GB of system memory in a single VM, receiving up to 1,000 teraflops of mixed precision hardware acceleration performance. Next-generation NVIDIA NVLink interconnects deliver up to 300 GB/s of GPU-to-GPU bandwidth, 9X over PCIe, boosting performance on deep learning and HPC workloads by up to 40%.

Visit the GPU documentation to stay up to date on the latest in pricing and regional availability for NVIDIA GPUs.

Try NVIDIA V100 on Compute Engine
NVIDIA GPUs on Google Kubernetes Engine

NVIDIA GPUs in Google Kubernetes Engine (GKE) are generally available and ready to be used widely from the latest GKE release.

Using GPUs in GKE can turbocharge compute-intensive applications like machine learning, image processing, and financial modeling. By packaging your CUDA workloads into containers, you can benefit from the massive processing power of GKE’s NVIDIA GPUs whenever you need it, without having to manage hardware or even VMs.

NVIDIA Tesla P4, V100, P100, and K80 GPUs are now generally available.

GPUs in Kubernetes Engine
NVIDIA GPU Cloud and Google Cloud Platform: GCP adds support for NVIDIA GPU Cloud

Google Cloud Platform has now added support for NVIDIA GPU Cloud. NVIDIA GPU Cloud (NGC) provides simple access to GPU-accelerated software containers for deep learning, HPC applications, and HPC visualization. NGC containers are optimized and pre-integrated to run GPU-accelerated software that takes full advantage of NVIDIA Tesla V100 and P100 GPUs on Google Cloud Platform. By making it simple to access NVIDIA GPU-accelerated software and NVIDIA GPUs, Google Cloud Platform is now helping you deploy production-quality GPU-optimized software in just minutes.

Get Started with NVIDIA CPU Cloud Image on Google Cloud Platform Marketplace

Additional Resources