Google upgraded Gemini with Gemini 2.5 Flash Image, a native image capability enabling finer, natural-language-driven photo edits. The model performs precise edits while preserving faces, animals, and other scene details that many rival tools distort during editing. The rollout begins Tuesday across the Gemini app and developer platforms including the Gemini API, Google AI Studio, and Vertex AI. The model surfaced anonymously on LMArena as "nano-banana" and scored strongly on benchmarks. Google describes the update as pushing visual quality and instruction-following, making edits more seamless and producing outputs usable for a range of purposes.
Google is upgrading its Gemini chatbot with a new AI image model that gives users finer control over editing photos, a step meant to catch up with OpenAI's popular image tools and draw users from ChatGPT. The update, called Gemini 2.5 Flash Image, rolls out starting Tuesday to all users in the Gemini app, as well as to developers via the Gemini API, Google AI Studio, and Vertex AI platforms.
Gemini's new AI image model is designed to make more precise edits to images - based on natural language requests from users - while preserving the consistency of faces, animals, and other details, something that most rival tools struggle with. For instance, ask ChatGPT or xAI's Grok to change the color of someone's shirt in a photo, and the result might include a distorted face or an altered background.
Google's new tool has already drawn attention. In recent weeks, social media users raved over an impressive AI image editor in the crowdsourced evaluation platform, LMArena. The model appeared to users anonymously under the pseudonym "nano-banana." Google says it's behind the model (if it wasn't obvious already from all the banana-related hints), which is really the native image capability within its flagship Gemini 2.5 Flash AI model. Google says the image model is state-of-the-art on LMArena and other benchmarks.
Collection
[
|
...
]