Read more at source.
Read more at source.
Google is also highlighting Gemini's native multimodality as a key advantage. The model is capable of interpreting not just text, but also audio, still images, video, and code. In addition, Google has announced that a 2 million token context window is soon to be introduced to help the model process more data.
Google's Gemini models have evolved into reasoning models that process tasks step-by-step and make more informed decisions. This results in better answers and responses for complex prompts. Google is now integrating these thinking capabilities directly into all of its models, enabling them to handle more complex problems and support even more capable, context-aware agents.
Google DeepMind CEO Demis Hassabis has praised Gemini 2.5 Pro as an 'awesome state-of-the-art model', ranking first on LMArena by a significant +39 ELO points. The model has shown significant improvements across the board in multimodal reasoning, coding & STEM. A demo video shows 2.5 Pro using its reasoning capabilities to program a video game based on a single prompt.
Google says its jumped forward in quality because Gemini models are now reasoning models that process tasks step-by-step and make more informed decisions, which they say results in better answers and responses for complex prompts.