Build with our next generation AI systems

Explore models chevron_right

Gemini

Our most intelligent AI models

2.5 Pro
2.5 Flash
2.0 Flash-Lite
Learn more

Gemma

Lightweight, state-of-the-art open models

Gemma 3
Gemma 3n
ShieldGemma 2
Learn more

Generative models

Image, music and video generation models

Imagen
Lyria
Veo

Experiments

AI prototypes and experiments

Project Astra
Project Mariner
Gemini Diffusion

Our latest AI breakthroughs and updates from the lab

Explore research chevron_right

Projects

Explore some of the biggest AI innovations

Learn more

Publications

Read a selection of our recent papers

Learn more

News

Discover the latest updates from our lab

Learn more

Unlocking a new era of discovery with AI

Explore science chevron_right

AI for biology

AlphaFold
AlphaMissense
AlphaProteo

AI for climate and sustainability

WeatherNext

AI for mathematics and computer science

AlphaEvolve
AlphaProof
AlphaGeometry

AI for physics and chemistry

GNoME
Fusion
AlphaQubit

AI transparency

SynthID

Our mission is to build AI responsibly to benefit humanity

About Google DeepMind chevron_right

News

Discover our latest AI breakthroughs, projects, and updates

Learn more

Careers

We’re looking for people who want to make a real, positive impact on the world

Learn more

Milestones

For over 20 years, Google has worked to make AI helpful for everyone

Learn more

Education

We work to make AI more accessible to the next generation

Learn more

Responsibility

Ensuring AI safety through proactive security, even against evolving threats

Learn more

The Podcast

Uncover the extraordinary ways AI is transforming our world

Learn more
Models
Research
Science
About

Build with our next generation AI systems

Explore models chevron_right

Gemini

Our most intelligent AI models

2.5 Pro
2.5 Flash
2.0 Flash-Lite
Learn more

Gemma

Lightweight, state-of-the-art open models

Gemma 3
Gemma 3n
ShieldGemma 2
Learn more

Generative models

Image, music and video generation models

Imagen
Lyria
Veo

Experiments

AI prototypes and experiments

Project Astra
Project Mariner
Gemini Diffusion

Our latest AI breakthroughs and updates from the lab

Explore research chevron_right

Projects

Explore some of the biggest AI innovations

Learn more

Publications

Read a selection of our recent papers

Learn more

News

Discover the latest updates from our lab

Learn more

Unlocking a new era of discovery with AI

Explore science chevron_right

AI for biology

AlphaFold
AlphaMissense
AlphaProteo

AI for climate and sustainability

WeatherNext

AI for mathematics and computer science

AlphaEvolve
AlphaProof
AlphaGeometry

AI for physics and chemistry

GNoME
Fusion
AlphaQubit

AI transparency

SynthID

Our mission is to build AI responsibly to benefit humanity

About Google DeepMind chevron_right

News

Discover our latest AI breakthroughs, projects, and updates

Learn more

Careers

We’re looking for people who want to make a real, positive impact on the world

Learn more

Milestones

For over 20 years, Google has worked to make AI helpful for everyone

Learn more

Education

We work to make AI more accessible to the next generation

Learn more

Responsibility

Ensuring AI safety through proactive security, even against evolving threats

Learn more

The Podcast

Uncover the extraordinary ways AI is transforming our world

Learn more
Models

Build with our next generation AI systems

Explore models chevron_right

Gemini

Our most intelligent AI models

2.5 Pro
2.5 Flash
2.0 Flash-Lite
Learn more

Gemma

Lightweight, state-of-the-art open models

Gemma 3
Gemma 3n
ShieldGemma 2
Learn more

Generative models

Image, music and video generation models

Imagen
Lyria
Veo

Experiments

AI prototypes and experiments

Project Astra
Project Mariner
Gemini Diffusion
Research

Our latest AI breakthroughs and updates from the lab

Explore research chevron_right

Projects

Explore some of the biggest AI innovations

Learn more

Publications

Read a selection of our recent papers

Learn more

News

Discover the latest updates from our lab

Learn more
Science

Unlocking a new era of discovery with AI

Explore science chevron_right

AI for biology

AlphaFold
AlphaMissense
AlphaProteo

AI for climate and sustainability

WeatherNext

AI for mathematics and computer science

AlphaEvolve
AlphaProof
AlphaGeometry

AI for physics and chemistry

GNoME
Fusion
AlphaQubit

AI transparency

SynthID
About

Our mission is to build AI responsibly to benefit humanity

About Google DeepMind chevron_right

News

Discover our latest AI breakthroughs, projects, and updates

Learn more

Careers

We’re looking for people who want to make a real, positive impact on the world

Learn more

Milestones

For over 20 years, Google has worked to make AI helpful for everyone

Learn more

Education

We work to make AI more accessible to the next generation

Learn more

Responsibility

Ensuring AI safety through proactive security, even against evolving threats

Learn more

The Podcast

Uncover the extraordinary ways AI is transforming our world

Learn more
Try Google AI Studio Try Gemini
Google DeepMind
Google AI Learn about all of our AI Google DeepMind Explore the frontier of AI Google Labs Try our AI experiments Google Research Explore our research
Gemini app Chat with Gemini Google AI Studio Build with our next-gen AI models
Models Research Science About
Try Google AI Studio Try Gemini

Gemini

Our most intelligent AI models

Chat with Gemini Try in Google AI Studio

Models

Gemini 2.5: Our most intelligent models are getting even better

Models

Gemini 2.5 Pro Preview: even better coding performance

Models

Build rich, interactive web apps with an updated Gemini 2.5 Pro

Gemini 2.5 models are capable of reasoning through their thoughts before responding, resulting in enhanced performance and improved accuracy.

  • What's new
  • Models
  • Hands-on
  • Performance
  • Safety
  • Build

What's new

Access the latest preview of Gemini 2.5 Pro

We’re introducing an upgraded preview of Gemini 2.5 Pro, our most intelligent model yet.

Try in Google AI Studio

Deep Think

We’re making Gemini 2.5 Pro even better by introducing an enhanced reasoning mode called Deep Think.

Learn more

Native audio

Converse in more expressive ways with native audio outputs that capture the subtle nuances of how we speak.

Learn more

An even better 2.5 Flash

Improved across key benchmarks for reasoning, multimodality, code and long context while getting even more efficient.

Try in Google AI Studio

Model family

Gemini 2.5 builds on the best of Gemini — with native multimodality and a long context window.

  • Preview

    2.5 Pro

    Best for coding and complex prompts

    Learn more
  • Preview

    2.5 Flash

    Best for fast performance on complex tasks

    Learn more
  • General availability

    2.0 Flash-Lite

    Best for cost-efficient performance

    Learn more

Hands-on with 2.5 Pro

See how Gemini 2.5 Pro uses its reasoning capabilities to create interactive simulations and do advanced coding.

Make an interactive animation

See how Gemini 2.5 Pro uses its reasoning capabilities to create an interactive animation of “cosmic fish” with a simple prompt.

Create your own dinosaur game

Watch Gemini 2.5 Pro create an endless runner game, using executable code from a single line prompt.

Code a fractal visualization

See how Gemini 2.5 Pro creates a simulation of intricate fractal patterns to explore a Mandelbrot set.

Plot interactive economic data

Watch Gemini 2.5 Pro use its reasoning capabilities to create an interactive bubble chart to visualize economic and health indicators over time.

Animate complex behavior

See how Gemini 2.5 Pro creates an interactive Javascript animation of colorful boids inside a spinning hexagon.

Code particle simulations

Watch Gemini 2.5 Pro use its reasoning capabilities to create an interactive simulation of a reflection nebula.

Performance

Gemini 2.5 is state-of-the-art across a wide range of benchmarks.

Benchmarks

Gemini 2.5 Pro demonstrates significantly improved performance across a wide range of benchmarks.

Benchmark
Gemini 2.5 Pro Preview 06-05
Thinking
OpenAI o3 High
OpenAI o4-mini High
Claude Opus 4 32k thinking
Grok 3 Beta Extended thinking
DeepSeek R1 05-28
Input price
$/1M tokens
(no caching)
$1.25
$2.50 > 200k tokens
$10.00 $1.10 $15.00 $3.00 $0.55
Output price
$/1M tokens $10.00
$15.00 > 200k tokens
$40.00 $4.40 $75.00 $15.00 $2.19
Reasoning & knowledge Humanity's Last Exam (no tools)
21.6% 20.3% 14.3% 10.7% — 14.0%*
Science GPQA diamond
single attempt 86.4% 83.3% 81.4% 79.6% 80.2% 81.0%
   
multiple attempts — — — 83.3% 84.6% —
Mathematics AIME 2025
single attempt 88.0% 88.9% 92.7% 75.5% 77.3% 87.5%
   
multiple attempts — — — 90.0% 93.3% —
Code generation LiveCodeBench (UI: 1/1/2025-5/1/2025)
single attempt 69.0% 72.0% 75.8% 51.1% — 70.5%
Code editing Aider Polyglot
82.2% diff-fenced
79.6% diff
72.0% diff
72.0% diff
53.3% diff
71.6%
Agentic coding SWE-bench Verified
single attempt 59.6% 69.1% 68.1% 72.5% — —
   
multiple attempts 67.2% — — 79.4% — 57.6%
Factuality SimpleQA
54.0% 48.6% 19.3% — 43.6% 27.8%
Factuality FACTS grounding
87.8% 69.6% 62.1% 77.7% 74.8% —
Visual reasoning MMMU
single attempt 82.0% 82.9% 81.6% 76.5% 76.0% no MM support
   
multiple attempts — — — — 78.0% no MM support
Image understanding Vibe-Eval (Reka)
67.2% — — — — no MM support
Video understanding VideoMMMU
83.6% — — — — no MM support
Long context MRCR v2 (8-needle)
128k (average) 58.0% 57.1% 36.3% — 34.0% —
   
1M (pointwise) 16.4% no support no support no support no support no support
Multilingual performance Global MMLU (Lite)
89.2% — — — — —

Methodology

Gemini results: All Gemini scores are pass @1."Single attempt" settings allow no majority voting or parallel test-time compute; "multiple attempts" settings allow test-time selection of the candidate answer. They are all run with the AI Studio API for the model-id gemini-2.5-pro-preview-06-05 with default sampling settings. To reduce variance, we average over multiple trials for smaller benchmarks. Aider Polyglot score is the pass rate average of 3 trials. Vibe-Eval results are reported using Gemini as a judge.

Non-Gemini results: All the results for non-Gemini models are sourced from providers' self reported numbers unless mentioned otherwise below.

All SWE-bench Verified numbers follow official provider reports, using different scaffoldings and infrastructure. Google's scaffolding for "multiple attempts" for SWE-Bench includes drawing multiple trajectories and re-scoring them using model's own judgement.

Thinking vs not-thinking: For Claude 4 results are reported for the reasoning model where available (HLE, LCB, Aider). For Grok-3 all results come with extended reasoning except for SimpleQA (based on xAI reports) and Aider. For OpenAI models high level of reasoning is shown where results are available (except for GPQA, AIME 2025, SWE-Bench, FACTS, MMMU).

Single attempt vs multiple attempts: When two numbers are reported for the same eval higher number uses majority voting with n=64 for Grok models and internal scoring with parallel test time compute for Anthropic models.

Result sources: Where provider numbers are not available we report numbers from leaderboards reporting results on these benchmarks: Humanity's Last Exam results are sourced from https://agi.safe.ai/ and https://scale.com/leaderboard/humanitys_last_exam, AIME 2025 numbers are sourced from https://matharena.ai/. LiveCodeBench results are from https://livecodebench.github.io/leaderboard.html (1/1/2025 - 5/1/2025 in the UI), Aider Polyglot numbers come from https://aider.chat/docs/leaderboards/. FACTS come from https://www.kaggle.com/benchmarks/google/facts-grounding. For MRCR v2 which is not publicly available yet we include 128k results as a cumulative score to ensure they can be comparable with other models and a pointwise value for 1M context window to show the capability of the model at full length. The methodology has changed in this table vs previously published results for MRCR v2 as we have decided to focus on a harder, 8-needle version of the benchmark going forward.

API costs are sourced from providers' website and are current as of June 5th.

* indicates evaluated on text problems only (without images)

Building responsibly in the agentic era

As we develop these new technologies, we recognize the responsibility it entails, and aim to prioritize safety and security in all our efforts.

Learn more

For developers

Gemini’s advanced thinking, native multimodality and massive context window empowers developers to build next-generation experiences.

Start building

Developer ecosystem

Build with cutting-edge generative AI models and tools to make AI helpful for everyone.

Google AI Studio

Build with the latest models from Google DeepMind

Gemini API

Easily integrate Google’s most capable AI model to your apps

Accessing our latest AI models

We want developers to gain access to our models as quickly as possible. We’re making these available through Google AI Studio.

Sign in to Google AI Studio

Get the latest updates

Sign up for news on the latest innovations from Google DeepMind.

Please enter a valid email (e.g., "[email protected]")

I accept Google's Terms and Conditions and acknowledge that my information will be used in accordance with Google's Privacy Policy.

Follow us
footer__x
footer__instagram
footer__youtube
footer__linkedin
footer__github
Build AI responsibly to benefit humanity
Models
Build with our next generation AI systems
Gemini
Gemma
Veo
Imagen
Lyria
Science
Unlocking a new era of discovery with AI
AlphaFold
SynthID
WeatherNext
Learn more
About News Careers Research Responsibility & Safety
Sign up for updates on our latest innovations

I accept Google's Terms and Conditions and acknowledge that my information will be used in accordance with Google's Privacy Policy.

Please enter a valid email (e.g., "[email protected]")
About Google
Google products
Privacy
Terms
Gemini Flash
Gemini Flash

Preview

Gemini 2.5 Flash

Best for fast performance on complex tasks

Try in Google AI Studio

Our powerful and most efficient workhorse model designed for speed and low-cost.

Speed and value at scale

Ideal for tasks like summarization, chat applications, data extraction, and captioning.

  • Thinking budget

    Control how much 2.5 Flash reasons to balance latency and cost.

  • Natively multimodal

    Understands input across text, audio, images and video.

  • Long context

    Explore vast datasets with a 1-million token context window.

Adaptive and budgeted thinking

Adaptive controls and adjustable thinking budgets allow you to balance performance and cost.

  • Calibrated

    The model explores diverse thinking strategies, leading to more accurate and relevant outputs.

  • Controllable

    Developers have fine-grained control over the model's thinking process, allowing them to manage resource usage.

  • Adaptive

    When no thinking budget is set, the model assesses the complexity of a task and calibrates the amount of thinking accordingly.

Preview

Native audio

Converse in more expressive ways with native audio outputs that capture the subtle nuances of how we speak. Seamlessly switch between 24 languages, all with the same voice.

Try in Google AI Studio

Benchmarks

Benchmark
Gemini 2.5 Flash Preview (05-20)
Thinking
Gemini 2.0 Flash
OpenAI o4-mini
Claude 3.7 Sonnet 64k Extended thinking
Grok 3 Beta Extended thinking
DeepSeek R1
Input price
$/1M tokens $0.15 $0.10 $1.10 $3.00 $3.00 $0.55
Output price
$/1M tokens
$0.60
no thinking
$3.50
thinking
$0.40 $4.40 $15.00 $15.00 $2.19
Reasoning & knowledge Humanity's Last Exam (no tools)
11.0% 5.1% 14.3% 8.9% — 8.6%*
Science GPQA diamond
single attempt (pass@1) 82.8% 60.1% 81.4% 78.2% 80.2% 71.5%
   
multiple attempts — — — 84.8% 84.6% —
Mathematics AIME 2025
single attempt (pass@1) 72.0% 27.5% 92.7% 49.5% 77.3% 70.0%
   
multiple attempts — — — — 93.3% —
Code generation LiveCodeBench v5
single attempt (pass@1) 63.9% 34.5% — — 70.6% 64.3%
   
multiple attempts — — — — 79.4% —
Code editing Aider Polyglot
61.9% / 56.7% whole / diff-fenced
22.2% whole
68.9% / 58.2% whole / diff
64.9% diff
53.3% diff
56.9% diff
Agentic coding SWE-bench Verified
60.4% — 68.1% 70.3% — 49.2%
Factuality SimpleQA
26.9% 29.9% — — 43.6% 30.1%
Factuality FACTS Grounding
85.3% 84.6% 62.1% 78.8% 74.8% 56.8%
Visual reasoning MMMU
single attempt (pass@1) 79.7% 71.7% 81.6% 75.0% 76.0% no MM support
   
multiple attempts — — — — 78.0% no MM support
Image understanding Vibe-Eval (Reka)
65.4% 56.4% — — — no MM support
Long context MRCR v2
128k (average) 74.0% 36.0% 49.0% — 54.0% 45.0%
   
1M (pointwise) 32.0% 6.0% — — — —
Multilingual performance Global MMLU (Lite)
88.4% 83.4% — — — —

Methodology

Gemini results: All Gemini scores are pass @1 (no majority voting or parallel test time compute unless indicated otherwise). They are all run with the AI Studio API for the model-id gemini-2.5-flash-preview-05-20 and gemini-2.0-flash with default sampling settings. To reduce variance, we average over multiple trials for smaller benchmarks. Vibe-Eval results are reported using Gemini as a judge.

Non-Gemini results: All the results for non-Gemini models are sourced from providers' self reported numbers unless mentioned otherwise below. All SWE-bench Verified numbers follow official provider reports, using different scaffoldings and infrastructure. Google's scaffolding includes drawing multiple trajectories and re-scoring them using model's own judgement.

Thinking vs not-thinking: For Claude 3.7 Sonnet: GPQA, AIME 2024, MMMU come with 64k extended thinking, Aider with 32k, and HLE with 16k. Remaining results come from the non thinking model due to result availability. For Grok-3 all results come with extended reasoning except for SimpleQA (based on xAI reports) and Aider.

Single attempt vs multiple attempts: When two numbers are reported for the same eval higher number uses majority voting with n=64 for Grok models and internal scoring with parallel test time compute for Anthropic models.

Result sources: Where provider numbers are not available we report numbers from leaderboards reporting results on these benchmarks: Humanity's Last Exam results are sourced from https://agi.safe.ai/ and https://scale.com/leaderboard/humanitys_last_exam, AIME 2025 numbers are sourced from https://matharena.ai/. LiveCodeBench results are from https://livecodebench.github.io/leaderboard.html (10/1/2024 - 2/1/2025 in the UI), Aider Polyglot numbers come from https://aider.chat/docs/leaderboards/. FACTS come from https://www.kaggle.com/benchmarks/google/facts-grounding. For MRCR v2 which is not publically available yet we include 128k results as a cumulative score to ensure they can be comparable with previous results and a pointwise value for 1M context window to show the capability of the model at full length.

API costs are sourced from providers' website and are current as of May 20th.

* indicates evaluated on text problems only (without images)

Model information

2.5 Flash Model Card

2.0 Flash 2.5 Flash
Model deployment status General availability Preview
Supported data types for input Text, Image, Video, Audio Text, Image, Video, Audio
Supported data types for output Text Text
Supported # tokens for input 1M 1M
Supported # tokens for output 8k 64k
Knowledge cutoff June 2024 January 2025
Tool use Search as a tool
Code execution
Function calling
Structured output
Search as a tool
Code execution
Best for Low latency scenarios
Automating tasks
Cost-efficient thinking
Well-rounded capabilities
Availability Google AI Studio
Gemini API
Vertex AI
Gemini App
Google AI Studio
Gemini API
Vertex AI
Gemini App

Try Gemini Flash

Preview

2.5 Flash

Best for fast performance on complex tasks

General availability

2.0 Flash

Best for fast performance on everyday tasks

General availability

2.0 Flash-Lite

Best for cost-efficient performance