Back to stories
Models

Google Launches Gemini 3.1 Flash Live and Search Live Globally — Point Your Camera and Talk

Michael Ouroumis3 min read
Google Launches Gemini 3.1 Flash Live and Search Live Globally — Point Your Camera and Talk

Google has launched Gemini 3.1 Flash Live, its highest-quality conversational voice model yet, and used it to power a global rollout of Search Live across more than 200 countries.

The capability is striking in its simplicity: point your phone camera at anything — a plant, a dish, a street sign, a broken appliance — and have a real-time spoken conversation about it. Not a text query. Not a search result page. A live dialogue.

What's New in Gemini 3.1 Flash Live

Gemini 3.1 Flash Live is optimized for real-time voice interaction — low latency, high naturalness, and the ability to maintain conversational context across a back-and-forth exchange. It's the model powering the voice layer of Search Live, and it represents a significant step up from previous Gemini voice capabilities in terms of both response speed and conversational coherence.

The "Flash" designation indicates it's tuned for speed and efficiency rather than maximum reasoning depth — the right trade-off for a voice-first, real-time product where sub-second response times matter more than extended reasoning chains.

Search Live Goes Global

Search Live — Google's camera-plus-voice search feature — has been in limited testing for months. Today's global rollout to 200+ countries is the moment it becomes a mainstream product rather than an experiment.

The user experience is designed to be frictionless. Open the camera, point it at something, ask a question out loud. Follow up conversationally. Ask for more detail, a different angle, a recommendation. The AI maintains context across the exchange so you don't have to restart each question from scratch.

This is a direct evolution of Google Lens, which was already the most-used image-based search product globally. Adding real-time voice conversation to visual search creates something qualitatively different from any search experience that came before.

The Gemini Drop: Memory Portability

Alongside the Flash Live launch, Google's March Gemini Drop included a feature that's strategically significant: the ability to import AI memories and full chat history from rival apps — including ChatGPT — directly into Gemini.

The move is a direct play for user retention and switching. One of the stickiest parts of any AI assistant is accumulated context — the model knowing your preferences, your past projects, your communication style. Google is removing that as a barrier to switching, betting that users who try Gemini will stay once they see its capabilities, rather than being locked out by the migration friction.

Combined with Apple's announcement that iOS 27 will open Siri to rival AIs including Gemini, Google is executing a coordinated distribution push: be everywhere users are, make it easy to start using Gemini, and demonstrate quality once they arrive.

What This Means

The voice AI layer is becoming the primary battleground. Mistral shipped Voxtral TTS earlier today. Cohere shipped Transcribe. Apple is opening Siri to Gemini. And Google is deploying its best voice model into the world's most-used search product.

The era of text-primary AI interaction is not ending — but it's sharing the stage with voice in a way it hasn't before. The companies winning the voice interface layer will have distribution advantages that compound: users who start asking questions by speaking rather than typing will default to whichever voice experience feels most natural. Google just made a strong claim on being that default.

Learn AI for Free — FreeAcademy.ai

Take "AI Essentials: Understanding AI in 2026" — a free course with certificate to master the skills behind this story.

More in Models

OpenAI Ships GPT-Realtime-2 With Live Translation and Streaming Whisper, Pushing Voice Agents Toward GPT-5 Reasoning
Models

OpenAI Ships GPT-Realtime-2 With Live Translation and Streaming Whisper, Pushing Voice Agents Toward GPT-5 Reasoning

OpenAI launched three new audio models on May 7 — GPT-Realtime-2, GPT-Realtime-Translate, and GPT-Realtime-Whisper — adding GPT-5-class reasoning, a 128K context window, and metered live translation across 70 languages to its already-GA Realtime API.

3 days ago2 min read
Zyphra Releases ZAYA1-8B, the First Frontier-Class Reasoning MoE Trained Entirely on AMD
Models

Zyphra Releases ZAYA1-8B, the First Frontier-Class Reasoning MoE Trained Entirely on AMD

Zyphra's ZAYA1-8B is an Apache-licensed reasoning MoE with under 1B active parameters that matches Claude 4.5 Sonnet and Gemini 2.5 Pro on math benchmarks — and it was pretrained, midtrained, and fine-tuned on AMD Instinct MI300X GPUs.

3 days ago2 min read
OpenAI Makes GPT-5.5 Instant the Default ChatGPT Model With 52.5% Fewer Hallucinations
Models

OpenAI Makes GPT-5.5 Instant the Default ChatGPT Model With 52.5% Fewer Hallucinations

OpenAI replaced GPT-5.3 Instant with GPT-5.5 Instant as the default ChatGPT model, citing a 52.5% reduction in hallucinated claims on high-stakes prompts and major gains on math and multimodal benchmarks.

5 days ago2 min read