The world of generative AI is moving fast, but sometimes the most powerful breakthroughs arrive with the most ridiculous names. That’s the story of Google’s latest viral image generation model, officially known as Gemini 2.5 Flash Image, but universally recognized by its accidental moniker: Nano Banana.
In a revealing episode of the Made by Google Podcast, David Chiron, a Group Product Manager on the Gemini app team, pulled back the curtain on the model’s meteoric rise, its technological leap, and how a late-night joke became a global brand.
The Midnight Name Drop
The story of “Nano Banana” is one of pure serendipity. As David Chiron explained, the name was never intended for public consumption. When the Google team submitted their new model anonymously to LM Arena—an external website that ranks different AI models based on user votes—they needed a simple, immediate placeholder.
“The official name is… much more catchy, Gemini 2.5 Flash Image,” Chiron joked. The truth, however, is that the nickname was born around 2:30 AM when a Product Manager named Nina had “a moment of brilliance” and typed in “Nano Banana.” The team expected the model to perform well, but they never anticipated the placeholder name would stick. When the model topped the charts and began trending on X (formerly Twitter), the public had already chosen its favorite fruit-themed title. Google decided to embrace the accidental branding, even adding the banana emoji to the Gemini app to indicate where the model’s new capabilities are in use.
The Breakthrough of “No More AI Distant Cousins”
The model’s viral success is underpinned by a crucial technological leap: unparalleled character and facial consistency. For David Chiron, the turning point was a personal experiment.
“The first time I tried Nano Banana, I uploaded an image of myself and asked to put myself in space,” he recounted. “All of a sudden, for the first time, I saw myself in the image and not my AI distant cousin.”
This precision—the ability to render a user’s face, pet, or loved one accurately across different images and scenarios—is what makes the model “big.” As humans, we are highly trained to spot tiny imperfections in faces, from a misplaced wrinkle to an awkward smile. Gemini 2.5 Flash Image solved the challenge of maintaining that high level of detail and identity, allowing people to truly imagine themselves and others in new, impossible ways.
From Figurines to Emotional Tributes
Since its launch, the “Nano Banana” model has generated billions of images, spawning global viral trends and facilitating deeply personal moments.
One of the first major trends was the figurine trend, which originated in Thailand. Users discovered a highly specific, nearly 90-word prompt that consistently produced a popular style of collectible figurine featuring their likeness. This trend quickly hopped across Asia, South America (Mexico and Brazil), and eventually went global.
Beyond the viral fun, the model has enabled profoundly emotional uses:
- The Polaroid Trend: Users utilized the model to create lifelike Polaroid images, often bringing together loved ones who could not be present for major life events. Stories emerged of users creating photos of a baby with a deceased grandmother for a baby shower, or a graduate including their deceased father in their graduation picture.
- Image Restoration: The model is also being used to restore family history. Users can upload old, black-and-white photos that are torn, wrinkled, or water-stained, and the AI can completely restore and even colorize them, effectively “reliving that moment” with a modern digital fidelity.
Responsibility in a New Era of Media
With such powerful image generation comes heavy responsibility, a fact Google’s team is keenly aware of. As David Chiron noted, the company operates under a philosophy of being both “bold and responsible,” honoring user requests while managing the societal impact.
To address concerns about digital provenance—knowing if an image is real—Google has implemented two layers of watermarking: - Visible Watermark: A clear watermark placed on the bottom right side of every image generated by the Gemini app, indicating it is AI-generated.
- Synth ID (Invisible Watermark): An unbreakable, invisible watermark that is imperceptible to the human eye. Because it is tied to the image’s deep code, it remains intact even if the image is cropped, edited, or compressed. Google can use this technology to detect if an image that gains prominence was generated by their model.
The Future: Listening and Video Generation
Looking ahead, the team is focusing on what they do best: listening. All feedback, from thumbs-up clicks to online complaints, is aggregated to guide the next round of improvements.
Chiron also highlighted related advancements in video generation with the recent update to the V3.1 model. This update brings major quality enhancements, specifically for photo-to-video (starting a video with a reference photo) and reference-to-video (placing a consistent object or likeness anywhere in a generated video). These tools offer a direct path for users to take their stunning “Nano Banana” images and bring them to life.
For those eager to try out the Gemini 2.5 Flash Image model, the recommendation is simple: start with one of the built-in templates, upload a photo, and create your own instant viral figurine—or, perhaps, your own accidental name.