Google AI

Gemini 2.0 Flash Thinking

Preview frontier

Experimental reasoning model with transparent thought process

Released December 19, 2024

Context Window
1M
TTFT
N/A
Speed
N/A
Max Output
64K
Training Cutoff
Aug 2024

Last tested: 2026-01-15

About

Gemini 2.0 Flash Thinking is an experimental model that exposes its reasoning process through visible 'thinking' tokens. Built on the Gemini 2.0 Flash architecture, it excels at complex multi-step problems requiring mathematical reasoning, coding challenges, and scientific analysis while providing interpretable chain-of-thought explanations.

Capabilities

vision function-calling streaming code-execution reasoning

Pricing

Input
$0.00/M
per 1M tokens
Output
$0.00/M
per 1M tokens

Static pricing

Details

Overview

Gemini 2.0 Flash Thinking is Google’s answer to the growing demand for transparent AI reasoning. This experimental model surfaces its internal thought process, allowing developers to understand how the model arrives at its conclusions. Currently free during the experimental phase, it represents Google’s investment in interpretable AI.

Key Features

Visible Reasoning Process

The model outputs explicit “thinking” tokens that reveal its step-by-step reasoning, making it easier to debug, validate, and trust AI-generated solutions.

Extended Output Length

With a 64K token output limit, Gemini 2.0 Flash Thinking can work through complex, multi-step problems without truncation, producing thorough analyses and detailed solutions.

Strong Performance on Hard Problems

Optimized for challenges that require careful reasoning:

  • Advanced mathematics and proof construction
  • Complex coding problems and debugging
  • Scientific analysis and hypothesis testing
  • Multi-step logical deduction

Performance Characteristics

While slower than the standard Flash model due to the reasoning overhead, Flash Thinking delivers significantly higher accuracy on challenging benchmarks:

  • AIME 2024: Strong performance on competitive math problems
  • GPQA: Improved scores on graduate-level science questions
  • Complex Coding: Better success rates on multi-file programming tasks

Use Cases

  • Educational Tools: Show students how to approach complex problems
  • Code Review: Understand the reasoning behind suggested code changes
  • Research Assistance: Get transparent analysis of scientific questions
  • Decision Support: Follow the model’s logic for critical decisions

Limitations

  • Currently in experimental phase with usage restrictions
  • Slower response times due to reasoning overhead
  • May produce verbose outputs for simple questions
  • Thinking tokens count toward output token limits
Newsletter

Get the signal, skip the noise.

Weekly digest of new models and provider updates across 41+ compute providers. Curated for AI builders who ship.

New model releases
Capability updates
Provider status
bots.so
The AI Inference Model Index
© bots.so — The AI Inference Model Index

bots.so aggregates publicly available model deployment information from official provider sources. We are not affiliated with any model provider. Model availability changes rapidly; always verify on official sites.