Google announces faster, more efficient Gemini AI model

INSUBCONTINENT EXCLUSIVE:
We recently spoke with Google's Tulsee Doshi, who noted that the 2.5 Pro (Experimental) release was still prone to "overthinking" its
responses to simple queries
However, the plan was to further improve dynamic thinking for the final release, and the team also hoped to give developers more control
over the feature
That appears to be happening with Gemini 2.5 Flash, which includes "dynamic and controllable reasoning."The newest Gemini models will choose
a "thinking budget" based on the complexity of the prompt
This helps reduce wait times and processing for 2.5 Flash
Developers even get granular control over the budget to lower costs and speed things along where appropriate
Gemini 2.5 models are also getting supervised tuning and context caching for Vertex AI in the coming weeks.In addition to the arrival of
Gemini 2.5 Flash, the larger Pro model has picked up a new gig
Google's largest Gemini model is now powering its Deep Research tool, which was previously running Gemini 2.0 Pro
Deep Research lets you explore a topic in greater detail simply by entering a prompt
The agent then goes out into the Internet to collect data and synthesize a lengthy report.Google says that the move to Gemini 2.5 has
boosted the accuracy and usefulness of Deep Research
The graphic above shows Google's alleged advantage compared to OpenAI's deep research tool
These stats are based on user evaluations (not synthetic benchmarks) and show a greater than 2-to-1 preference for Gemini 2.5 Pro
reports.Deep Research is available for limited use on non-paid accounts, but you won't get the latest model
Deep Research with 2.5 Pro is currently limited to Gemini Advanced subscribers
However, we expect before long that all models in the Gemini app will move to the 2.5 branch
With dynamic reasoning and new TPUs, Google could begin lowering the sky-high costs that have thus far made generative AI unprofitable.