AI Infrastructure Leaders Forum

Google Cloud AI Infrastructure Leaders Forum

Go beyond surface-level AI discussions. Deep dive into foundational tech for large-scale generative AI with Google's top engineers and industry vanguards. Exclusive, in-person.

Thank you for your interest. Registration is now closed.

September 25, 2025

12:30 PM-4:00 PM EDT

25 11th Ave

New York, NY 10011

Building the next generation of AI requires more than just breakthrough models; it demands advancements in the underlying infrastructure. As models grow in complexity and scale, the real challenge lies in creating powerful, efficient, and innovative systems capable of training and deploying them.


Join us at the AI Infrastructure Leaders Forum and gain the blueprint for building at the forefront of generative AI. Move beyond theory to dissect the core systems that power large-scale AI. Explore how to optimize your entire training stack, from custom silicon to the software that brings them together. 


You’ll have an opportunity to:

  • Go behind the scenes on real-world architectural decisions with the industry vanguards who built and scaled demanding AI workloads.
  • Learn how to solve critical systems-level challenges by engaging directly with Google's most distinguished engineers and Fellows.
  • Get a candid look at the future of AI infrastructure and the critical interplay between hardware, software, and compilers.

This exclusive forum is designed for the technical leaders shaping the future of AI, with deep-dive discussions tailored to your architectural and systems-level expertise.


Seats are limited and will be granted to the first 100 pre-registered guests upon their arrival. 


Agenda

12:30 PM

Registration and lunch

1:00 PM

What’s next in Google Cloud infrastructure for high-performance AI inference

As AI models transition from training to production, the critical challenge becomes delivering fast, reliable, and cost-effective inference at scale.


In this exclusive, private session, we will share our strategic roadmap for Google Cloud's inference-optimized infrastructure. Get an inside look at how we are engineering our hardware and software to tackle the unique demands of large-scale AI workloads. We will detail upcoming advancements designed to deliver ultra-low latency, maximize throughput, and provide breakthrough price-performance for your most critical inference workloads. This is an interactive discussion where you can ask questions and provide direct feedback to the product leaders. 

Due to the sensitive nature of the content, all attendees must sign a Non-Disclosure Agreement (NDA) unless already covered by a customer agreement.

2:00 PM

Break

2:15 PM

How Google Cloud customers are building a scalable AI infrastructure

Go behind the scenes with the technology leaders from innovative companies in this technical panel discussion.  We’ll explore real-world applications of cloud infrastructure for cutting-edge AI workloads. Companies will showcase their AI infrastructure learnings, and discuss the critical technical trade-offs and architectural decisions made to support and scale demanding AI workloads.

2:30 PM

Optimizing AI infrastructure at scale: a technical walkthrough by Google Cloud PMs

Delivering state-of-the-art AI requires an infrastructure that is both powerful and economically viable. This technical briefing is designed for engineering leaders focused on the performance and efficiency of AI inference at scale. Google Cloud Product Managers will demonstrate how our co-designed hardware and software solve critical challenges in throughput, latency, and cost. See the systems-level approach that abstracts complexity while providing the control needed to serve the most demanding models.

3:15 PM

Happy hour and networking reception
Google Cloud