Nano-Banana (Gemini 2.5 Flash Image): the AI model from Google that is surprising creatives

Nano-Banana (Gemini 2.5 Flash Image): the AI model from Google that is surprising creatives

Nano-Banana (Gemini 2.5 Flash Image): the AI model from Google that is surprising creatives

18 de ago. de 2025

Nando

CEO | FOUNDER

The world of artificial intelligence is full of surprises, and the most recent one has the curious name Nano-Banana. Without an official announcement or launch campaign, this model simply appeared on LM Arena, an open platform for testing and comparing AI models. Just a few hours of use were enough for the community to realize that there is something special there.

Image reproduction: Theo Richardson (@theodesigning on X)

While other models stumble on delicate tasks, such as maintaining character consistency or editing faces accurately, the Nano-Banana seems to leap ahead. Users report impressive results, especially in text-guided editing tasks, something that has been a weak point for many tools until now.

Where did the Nano-Banana come from?

This was the question that stirred forums and social networks until August 26. There was speculation that it could be a secret project from big techs, related to Qwen Image, or even the fruit of an independent lab.

The mystery surrounding the Nano-Banana has finally come to an end. On the morning of August 26, the official profile of Gemini on X confirmed that the model belongs to Google.

Before the confirmation, the buzz had already increased when Sundar Pichai, the company's CEO, published an enigmatic post with just three banana emojis 🍌🍌🍌, without further explanations.

The suspense surrounding the authorship further increased the community's interest. Now, with the official revelation, the Nano-Banana enters the list of Google's AI tools, reinforcing the big tech's strategy to expand its creative portfolio.

Why does the Nano-Banana stand out?

Unlike models like Midjourney or Stable Diffusion, known for creating images from scratch, the Nano-Banana seems to go beyond and has an additional focus: editing accurately. This ability to combine creation and editing is precisely what has impressed the community.

Users report that it is capable of:

  • Handling complex editing prompts with ease

  • Completing entire faces based on descriptions like "imagine the entire person's face and create it"

  • Adjusting fine details, such as lighting, textures, and features, so that the result does not "give away" the editing

This capability has generated bold comparisons: some claim it might herald a new generation of models more focused on post-production and intelligent retouching than on simple image creation.

Where to test the Nano-Banana (Gemini 2.5 Flash Image)?

Until recently, it was only accessible for experimental testing on LM Arena, where you could compare it with other models in blind rounds. This means that you assess images without knowing which model generated them, choosing the best ones, and the community was increasingly choosing Nano-Banana's results.

What was previously a mysterious model from LM Arena now has an official name: Gemini 2.5 Flash Image. This means that the former Nano-Banana is no longer limited to blind tests but can now be used practically by developers and companies.

How to access the Nano-Banana (Gemini 2.5 Flash Image):

  1. Google AI Studio: you can try it for free in build mode, creating and remixing editing and image generation apps with simple prompts.

  2. Gemini API: available for developers, priced at US$ 30 for 1 million output tokens (each image consumes ~1290 tokens, or about US$ 0.039 per image).

  3. Vertex AI: for companies already using Google’s cloud ecosystem.

  4. OpenRouter.ai: in partnership with Google, brings the model to a base of over 3 million developers.

  5. fal.ai: another partnership that makes the model accessible to an even larger community of generative media creators.

👉 To get started, visit Google AI Studio and test in build mode.

What are the new features of Gemini 2.5 Flash Image?

The former Nano-Banana didn't grab attention for nothing. Now officially launched as Gemini 2.5 Flash Image, the model arrives with a set of features that specifically target the toughest points of visual creation with AI:

  • Character consistency: one of the biggest challenges in image generation is maintaining the same character or object across different prompts. Gemini 2.5 solves this by allowing a single person to be placed in various environments, creating products from multiple angles or generating a sequence of images with consistent visual identity.

  • Localized editing via prompt: it is possible to make point transformations using natural language commands. Practical examples: blur only the background of an image, remove a stain from clothing, alter someone’s pose, or even colorize a black-and-white photo.

  • World knowledge: beyond aesthetics, the model understands context. This opens up possibilities for uses in education, informative design, and prototyping, like interpreting diagrams or creating interactive visual materials from sketches.

  • Merging multiple images: you can merge different inputs into one, such as inserting a product into a new scene, changing the style of an environment, or generating dynamic mockups for an entire catalog.

What makes the Nano-Banana special for image editing?

Most image generation models focus on creation from scratch, but fail when it comes to editing existing images. This is where the Nano-Banana draws attention.

It can interpret textual commands more accurately, offering edits that respect the visual coherence of the scene. Whether changing lighting details, adjusting facial features, or completing missing areas, the model shows results that do not appear artificial.

This differential makes many creatives see the Nano-Banana as a tool aimed not only at image invention but also at intelligent post-production.

One of the qualities of the Nano Banana is the way it manages motion blur: the product remains sharp and highlighted, while integrating harmoniously into the background.

The alignment between the character's style and the product creates an excellent result. What’s most impressive is how it manages to maintain the same atmosphere without altering the original aesthetics and style.

Image reproduction: Daria_Surkova (@Dari_Designs on X)

Can Nano-Banana complete or generate faces?

Yes! This is perhaps the most discussed feature by the community. The model has proven capable of completing partially visible faces or even generating entire features from text descriptions.

A simple prompt like “imagine the whole person's face and create it” has already been enough for users to report extremely natural results, with eyes, mouth, and expression reconstructed without the usual "strange artifacts" common in other AIs.

This places the Nano-Banana on a promising level for applications that require photorealism and consistency, something that has been hard to achieve in AI portraits until now.

X user @cannn064 shared an interesting result: “I edited this image asking: imagine the entire person's face and create it. On the left is my input and on the right is the image that nano banana generated.”

Image reproduction: Can Hi (@cannn064 on X)

For what types of tasks is the Nano-Banana most suitable?

Based on the initial tests conducted by the community, the Nano-Banana seems to shine in a few scenarios:

  • Text-based image editing: small or large transformations guided solely by descriptions.

  • Face generation and completion: reconstructing features, entire faces, or subtle expression adjustments.

  • Creative transformations: changes in style, atmosphere, or framing of existing images.

  • Technical details: refining lighting, textures, and correcting inconsistencies.

This versatility positions it as a tool that communicates with both artists and designers as well as professionals in photography, fashion, and advertising, who deal with demands for visual consistency in their daily work.

Image reproduction: Chetaslua (@chetaslua on X)

What to expect for the future of Nano-Banana in the Google ecosystem?

With the confirmation that Nano-Banana is Gemini 2.5 Flash Image, we no longer speak of rumors, but of a real Google product with a clear roadmap.

The model was launched in preview on Google AI Studio, Gemini API, and Vertex AI, with the promise of gaining stability in the coming weeks. This means that the future of the Nano-Banana is directly linked to the evolution of the Gemini ecosystem, which should bring constant updates in quality, speed, and integration with other modalities (text, audio, and video).

Among the short- and medium-term possibilities are:

  • Official stability: the model is still in preview but will be made stable soon.

  • Expanded integrations: it is expected to be more deeply incorporated into Google Workspace and creative workflows in the cloud via Vertex AI.

  • More creative control: features like character consistency, blending multiple images, and transformations guided by natural language should evolve based on creators' feedback.

  • Access expansion: partnerships with platforms like OpenRouter and fal.ai indicate that the model will become increasingly accessible to the global creative community.

  • Security and traceability: all generated images carry an invisible watermark SynthID, a trust layer that is set to become standard in future Google models.

In summary: if at first the Nano-Banana was just a fun enigma for the community, now, as Gemini 2.5 Flash Image, it firmly establishes itself as part of a bigger strategy by Google to lead visual creation with AI on a global scale.

👉 Also read: Human's favorite tools for creating images, videos, and upscaling with AI

Conclusion: Is Nano-Banana a passing hype or a creative revolution?

The Nano-Banana is the type of novelty that the community discovers before the press. Is it just a curious experiment... or the first step of a historical turning point in image editing with AI?

But now the mystery is over: we already know that the Nano-Banana is developed by Google. It may have emerged quietly, but it is already positioning itself as one of the most promising models in the market. Whether for accuracy in image editing or for the ability to generate entire faces naturally, the Nano-Banana shows that we are entering a new phase of visual creation.

And if you want to go beyond testing and master the tools that are already transforming the market, the AI Video Creator Pro ⚡️🎬 has arrived. The course has been updated with the latest in video creation with AI: tools, techniques, and processes that are changing the creative market.

There are 70+ hands-on lessons covering the entire process:

🔹 Script and visual narrative
🔹 Direction and acting analysis
🔹 Cinematic camera movements
🔹 Lipsync and realistic expressions
🔹 Texturing, editing, and professional finishing

All this within the Workflow Human®, the same one we use in the studio for real campaigns for major brands. 👉 Secure your spot in AI Video Creator Pro.

Get the latest news from the world of AI and the Market

Get the latest news from the world of AI and the Market

Every Thursday at 10 AM, in your email inbox.

MIDJORNEY

ChatGPT

Get the latest news from the world of AI and the Market

Get the latest news from the world of AI and the Market

Every Thursday at 10 AM, in your email inbox.

MIDJORNEY

ChatGPT

Get the latest news from the world of AI and the Market

Get the latest news from the world of AI and the Market

Every Thursday at 10 AM, in your email inbox.

MIDJORNEY

ChatGPT