Gemini 2.0 Flash Thinking Mode is an experimental model that's trained to generate the "thinking process" the model goes through as part of its response. As a result, Gemini 2.0 Flash Thinking Mode is capable of stronger reasoning capabilities in its responses than the base Gemini 2.0 Flash model.
Use Thinking Mode
Thinking Mode is available as an experimental model in Vertex AI.
To use Thinking Mode, select the gemini-2.0-flash-thinking-exp-1219
model
in the Model drop-down menu.
Thoughts
The model's thinking process is returned as the first element of the
content.parts
list that is created when the model generates the response. For
example, the following code prints out only the model's thinking process:
response = client.models.generate_content(
model='gemini-2.0-flash-thinking-exp-1219', contents='Solve 3*x^3-5*x=1'
)
Markdown(response.candidates[0].content.parts[0].text)
You can see more examples of how to use Thinking Mode in our Colab notebook.
Limitations
Thinking Mode is an experimental model and has the following limitations:
- 32k token input limit
- Text and image input only
- 8k token output limit
- Text only output
- No built-in tool usage like Search or code execution
What's next?
Try Thinking Mode for yourself with our Colab notebook, or open the Vertex AI console and try prompting the model for yourself.