Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. learn more
In its latest attempt to redefine the AI landscape, Google has announced Gemini 2.0 Flash thinka multimodal reasoning model capable of dealing with complex problems with speed and transparency.
In a post on the X social networkGoogle CEO Sundar Pichai wrote that it was: “Our most exciting model yet :)”
And on the developer documentationGoogle explains, “Thinking Mode has stronger reasoning abilities in its answers than the base Gemini Flash Model 2.0,” previously Google's latest and greatest, released just eight days ago.
The new model supports only 32,000 input signals (approx 50-60 pages of text) and can produce 8,000 marks per production answer. In a sidebar on Google AI Studio, the company says it's better for “multiple understanding, reasoning” and “coding.”
Full details of the model's training process, architecture, licensing and costs have not yet been released. Currently, it shows zero cost per token in Google AI Studio.
Accessible and clearer reasoning
Unlike competitive reasoning models o1 and o1 mini from OpenAIGemini 2.0 allows users to access the reasoning step-by-step through a drop-down menu, offering a clearer and clearer view of how the model reaches conclusions.

By allowing users to see how decisions are made, Gemini 2.0 addresses long-standing concerns about operating as a “black box,” and gives this model – terms Licensing remains unclear – to the same extent as other open source models offered by competitors.
My early simple tests of the model showed accurately and quickly (within one to three seconds) some questions that have been very difficult for other AI models, such as counting the number of Rs in the word “Strawberry.” (See screenshot above).
In another experiment, when comparing two decimal numbers (9.9 and 9.11), the model systematically broke the problem down into smaller steps, from analyzing whole numbers to comparing decimal places.
These results are backed by independent third-party research from LM Arenawho named Gemini 2.0 Flash Thinking as the main performance model across all LLM categories.
Native support for image uploading and analysis
In another development over the competing OpenAI o1 family, Gemini 2.0 Flash Thinking is designed to process images from the jump.
o1 launched as a text-only module, but has since expanded to include image analysis and file uploading. Both models can only return text, at this time.
Gemini 2.0 Flash Thinking also does not support basics with Google Search, or integration with other Google apps and external third-party tools, according to the developer documentation.
The multi-module capability of Gemini 2.0 Flash Thinking expands its possible use cases, allowing it to deal with situations that combine different types of data.
For example, in one test, the model solved a puzzle that required the analysis of textual and visual elements, showing its flexibility in integration and reasoning across formats.
Developers can leverage these features through Google AI Studio and Vertex AI, where the model is available for testing.
As the AI landscape becomes more competitive, Gemini 2.0 Flash Thinking could mark the beginning of a new era for problem-solving models. Its ability to handle different types of data, offer observable reasoning, and perform at scale positions it as a major contender in the AI reasoning market, competing with the OpenAI family and beyond.
Source link