Generative AI overview
Learn about building generative AI applications
Generative AI on Vertex AI
Gemini Quickstart
Choose infrastructure for your generative AI application
When to use generative AI
Develop a generative AI application
Code samples and sample applications
Model exploration and hosting
Google Cloud provides a set of state-of-the-art foundation models through Vertex AI, including Gemini. You can also deploy a third-party model to either Vertex AI Model Garden or self-host on GKE or Compute Engine.
Google Models on Vertex AI (Gemini, Imagen)
Other models in the Vertex AI Model Garden
Text generation models via HuggingFace
AI/ML orchestration on GKE
GPUs on Compute Engine
Prompt design and engineering
Prompt design is the process of authoring prompt and response pairs to give language models additional context and instructions. After you author prompts, you feed them to the model as a prompt dataset for pretraining. When a model serves predictions, it responds with your instructions built in.
Vertex AI Studio
Overview of Prompting Strategies
Prompt Gallery
Grounding and RAG
Grounding connects AI models to data sources to improve the accuracy of responses and reduce hallucinations. RAG, a common grounding technique, searches for relevant information and adds it to the model's prompt, ensuring output is based on facts and up-to-date information.
Vertex AI grounding
Ground with Google Search
Vector embeddings in AlloyDB
Cloud SQL and pgvector
Integrating BigQuery data into your LangChain application
Vector embeddings in Firestore
Vector embeddings in Memorystore (Redis)
Agents and function calling
Agents make it easy to design and integrate a conversational user interface into your mobile app, while function calling extends the capabilities of a model.
Vertex AI Agent Builder
Vertex AI Function calling
Model customization and training
Specialized tasks, such as training a language model on specific terminology, might require more training than you can do with prompt design or grounding alone. In that scenario, you can use model tuning to improve performance, or train your own model.