
We recently spoke with Google's Tulsee Doshi, who noted that the 2.5 Pro (Experimental) release was still prone to "overthinking" its responses to simple queries.
However, the plan was to further improve dynamic thinking for the final release, and the team also hoped to give developers more control over the feature.
That appears to be happening with Gemini 2.5 Flash, which includes "dynamic and controllable reasoning."The newest Gemini models will choose a "thinking budget" based on the complexity of the prompt.
This helps reduce wait times and processing for 2.5 Flash.
Developers even get granular control over the budget to lower costs and speed things along where appropriate.
Gemini 2.5 models are also getting supervised tuning and context caching for Vertex AI in the coming weeks.In addition to the arrival of Gemini 2.5 Flash, the larger Pro model has picked up a new gig.
Google's largest Gemini model is now powering its Deep Research tool, which was previously running Gemini 2.0 Pro.
Deep Research lets you explore a topic in greater detail simply by entering a prompt.
The agent then goes out into the Internet to collect data and synthesize a lengthy report.Google says that the move to Gemini 2.5 has boosted the accuracy and usefulness of Deep Research.
The graphic above shows Google's alleged advantage compared to OpenAI's deep research tool.
These stats are based on user evaluations (not synthetic benchmarks) and show a greater than 2-to-1 preference for Gemini 2.5 Pro reports.Deep Research is available for limited use on non-paid accounts, but you won't get the latest model.
Deep Research with 2.5 Pro is currently limited to Gemini Advanced subscribers.
However, we expect before long that all models in the Gemini app will move to the 2.5 branch.
With dynamic reasoning and new TPUs, Google could begin lowering the sky-high costs that have thus far made generative AI unprofitable.