Gemini Release Notes
Last updated: Apr 4, 2026
- Apr 2, 2026
- Date parsed from source:Apr 2, 2026
- First seen by Releasebot:Apr 4, 2026
Improved Gemini audio models for powerful voice interactions
Gemini adds updated 2.5 Flash Native Audio for live voice agents, improving complex workflows, instruction following, and natural conversations. It also brings live speech translation beta to Google Translate and expands access across Google AI Studio, Vertex AI, Gemini Live, and Search Live.
Earlier this week, we introduced greater control over audio generation with an upgrade to our Gemini 2.5 Pro and Flash Text-to-Speech models.
But generating expressive speech is only one side of the conversation. Today, we’re releasing an updated Gemini 2.5 Flash Native Audio for live voice agents. This update improves the model’s ability to handle complex workflows, navigate user instructions, and hold natural conversations.
Gemini 2.5 Flash Native Audio is now available across Google products including Google AI Studio, Vertex AI, and has also started rolling out in Gemini Live and Search Live, bringing the naturalness of native audio to Search Live for the first time. This means you can more effectively brainstorm live with Gemini, get real-time help in Search Live, or build the next generation of enterprise-ready customer service agents.
Beyond powering helpful agents, native audio unlocks new possibilities for global communication. We’re introducing live speech translation, a capability that enables streaming speech-to-speech translation for headphones. It preserves the speaker’s intonation, pacing and pitch. This beta experience is rolling out in the Google Translate app starting today.
Live Voice Agents
Gemini 2.5 Flash Native Audio is now enabling a wide spectrum of conversational experiences.
To enable the breadth of use cases across surfaces and products, we have improved Gemini 2.5 Native Audio in three key areas:
- Sharper function calling: We’ve improved the model's reliability when triggering external functions. It can now more accurately identify when to fetch real-time information during a conversation and seamlessly weave that data back into the audio response, without breaking the flow. On ComplexFuncBench Audio, an eval that captures multi-step function calling with various constraints, Gemini 2.5 Native Audio leads with a score of 71.5%.
- Robust instruction following: The model is now better at handling complex instructions resulting in higher user satisfaction on content completeness. With a 90% adherence rate to developer instructions (up from 84%), it delivers more reliable outputs.
- Smoother conversations: We’ve achieved significant gains in multi-turn conversation quality. Gemini 2.5 Flash Native Audio is able to retrieve context from previous turns more effectively, creating more cohesive conversations.
What customers are saying
Google Cloud customers are already using Gemini’s native audio capabilities to drive real business results, from mortgage processing to customer calls.
- “Users often forget they’re talking to AI within a minute of using Sidekick, and in some cases have thanked the bot after a long chat…New Live API AI capabilities offered through Gemini [2.5 Flash Native Audio] empower our merchants to win.” – David Wurtz, VP of Product, Shopify
- "By integrating the Gemini 2.5 Flash Native Audio model…we've significantly enhanced Mia's capabilities since launching in May 2025. This powerful combination has enabled us to generate over 14,000 loans for our broker partners." – Jason Bressler, Chief Technology Officer, United Wholesale Mortgage (UWM)
- “Working with the Gemini 2.5 Flash Native Audio model through Vertex AI allows Newo.ai AI Receptionists to achieve unmatched conversational intelligence ... .They can identify the main speaker even in noisy settings, switch languages mid-conversation, and sound remarkably natural and emotionally expressive.” – David Yang, Co-founder, Newo.ai
Live Speech Translation
Gemini now natively supports new live speech-to-speech translation capabilities designed to handle both continuous listening and two-way conversation.
With continuous listening, Gemini automatically translates speech in multiple languages into a single target language. This allows you to put headphones in and hear the world around you in your language.
For two-way conversation, Gemini’s live speech translation handles translation between two languages in real-time, automatically switching the output language based on who is speaking. For example, if you speak English and want to chat with a Hindi speaker, you’ll hear English translations in real-time in your headphones, while your phone broadcasts Hindi when you’re done speaking.
Gemini’s live speech translation has a number of key capabilities that help in the real world:
- Language coverage: Translates speech in over 70 languages and 2000 language pairs by combining Gemini model’s world knowledge and multilingual capabilities with its native audio capabilities
- Style transfer: Captures the nuance of human speech, preserving the speaker’s intonation, pacing and pitch so the translation sounds natural.
- Multilingual input: Understands multiple languages simultaneously in a single session, helping you follow multilingual conversations without needing to fiddle around with language settings.
- Auto detection: Identifies the spoken language and begins translation, so you don’t even need to know what language is being spoken to start translating.
- Noise robustness: Filters out ambient noise so you can converse comfortably even in loud, outdoor environments.
Starting today, you can try it in a new beta experience in the Google Translate app for real-time translation in your headphones by connecting them to your device and tapping “Live translate.” This experience is rolling out to all Android devices in the US, Mexico and India with support for iOS and more regions coming soon.
Based on feedback, we will continue to iterate on this experience and bring it to more Google products including the Gemini API in 2026.
Get started today
Start building voice agents today with Gemini 2.5 Flash Native Audio, now generally available on Vertex AI and as preview in the Gemini API. Try it out in Google AI Studio.
Gemini 2.5 Flash and 2.5 Pro text-to-speech models are also available via the Gemini API in Google AI Studio. Get started with the speech generation docs, explore the prompting guide, or check out the Gemini API Cookbook to get started.
Original source Report a problem - Mar 25, 2026
- Date parsed from source:Mar 25, 2026
- First seen by Releasebot:Apr 4, 2026
Lyria 3 Pro: Create longer tracks in more Google products
Gemini adds Lyria 3 Pro, unlocking longer music tracks with stronger structural awareness and more creative control. The update expands Lyria to Vertex AI, Google AI Studio, the Gemini app, Google Vids and ProducerAI, bringing advanced music generation to more creators and businesses.
Introducing Lyria 3 Pro, which unlocks longer tracks with structural awareness. We’re also bringing Lyria to more Google products and surfaces.
Last month, we introduced Lyria 3, featuring custom music generation designed to spark creative expression. Now, we’re bringing our most advanced music generation model to more Google products, and introducing Lyria 3 Pro. This advanced version allows the creation of tracks up to 3 minutes long, with customization and creative control. Lyria 3 Pro better understands musical composition, so you can now prompt for specific elements like intros, verses, choruses and bridges. It’s great for experimenting with different styles or generating songs with complex transitions.
Lyria 3 Pro Musical Overview
Providing new places to generate music
High-quality music generation should be accessible wherever creativity happens. Whether you are an app developer, a business or music professional, or a creator, these integrations allow you to use Lyria’s advanced musical awareness to scale your production.
Vertex AI: Lyria 3 Pro is now in public preview on Vertex AI for businesses who require on-demand audio at scale. It gives organizations the ability to scale high-fidelity production, from rapidly generating bespoke soundtracks for gaming to integrating into creative tools, music and video platforms.
Google AI Studio and the Gemini API: For developers building the next generation of creative tools, Lyria 3 provides improved musical awareness and structural coherence to offer creative flexibility. Lyria 3 Pro is now available alongside Lyria RealTime in AI Studio.
Google Vids: Vids is an AI-powered video creation app that anyone can use. With Lyria 3 and Lyria 3 Pro in Vids, you can add custom music that matches your style for everything from creative projects to marketing videos. This is rolling out to Google Workspace customers and Google AI Pro & Ultra subscribers starting this week.
Gemini app: Longer generations with Lyria 3 Pro are now available in the Gemini app, starting with paid subscribers. Lyria 3 Pro’s enhanced customization offers more space to experiment and play with longer tracks. So now, you can add more details to bring your full vision to life, or create personalized tracks for vlogs, podcasts or tutorial videos.
ProducerAI: We recently introduced ProducerAI, a collaborative music creation tool, built by musicians looking for new ways to enhance their creative process. With Lyria 3 Pro, ProducerAI offers an agentic experience designed to help artists, producers and songwriters at every level iterate on comprehensive songs. It’s available globally to free and paid subscribers.
Partnering with creatives
We have been developing our music generation tools responsibly and in close partnership with the industry to ensure AI serves as a tool for creative expression.
Through our Music AI Sandbox, we provide musicians, producers and songwriters with a suite of experimental tools designed to expand their creative horizons. The insights from this collaboration helped shape the development of Lyria 3.
We’re inviting artists to integrate AI into their workflows to make sure our technology helps the people who use it. Grammy-winning producer Yung Spielburg used Lyria in his composition and production process for the score of the Google DeepMind short film “Dear Upstairs Neighbors.” And we’re also collaborating with DJ and producer François K, who used Lyria in an iterative process to create a soon-to-be-released song.
“Having explored the vast landscape of music tools, the progress on Lyria 3 is incredible, especially the fidelity and musicality. The fashion in which I use generative AI tools never boils down to 'one-button-click' prompting. Instead it’s becoming a versatile part of my arsenal, allowing me to refine ideas with realism and precision.”
François K
Responsibility was foundational, and remains integral in the design and training of Lyria 3, using materials that YouTube and Google has a right to use under our terms of service, partner agreements, and applicable law. To protect original expression, Lyria 3 and Gemini do not mimic artists; if a prompt names a creator, the model takes that as broad inspiration. Additionally, we employ filters to check outputs against existing content and users must adhere to the Terms of Service and Gen AI prohibited use policies, which prohibit violating others' intellectual property and privacy rights. All Lyria 3 and Lyria 3 Pro outputs are embedded with SynthID, our imperceptible watermark for identifying Google AI-generated content.
Lyria 3 Pro is rolling out to professionals, developers, organizations and everyday creators to help craft high quality music generations.
Original source Report a problem All of your release notes in one feed
Join Releasebot and get updates from Google and hundreds of other software products.
- Mar 3, 2026
- Date parsed from source:Mar 3, 2026
- First seen by Releasebot:Apr 4, 2026
Gemini 3.1 Flash-Lite: Built for intelligence at scale
Gemini introduces 3.1 Flash-Lite, its fastest and most cost-efficient Gemini 3 series model, now rolling out in preview for developers and enterprises. It brings strong performance at low cost for high-volume workloads, with thinking levels and support in Gemini API, AI Studio, and Vertex AI.
Get best-in-class intelligence for your highest-volume workloads.
Today, we're introducing Gemini 3.1 Flash-Lite, our fastest and most cost-efficient Gemini 3 series model. Built for high-volume developer workloads at scale, 3.1 Flash-Lite delivers high quality for its price and model tier.
Starting today, 3.1 Flash-Lite is rolling out in preview to developers via the Gemini API in Google AI Studio and for enterprises via Vertex AI.
Cost-efficiency without compromise
Priced at just $0.25/1M input tokens and $1.50/1M output tokens, 3.1 Flash-Lite delivers enhanced performance at a fraction of the cost of larger models. It outperforms 2.5 Flash with a 2.5X faster Time to First Answer Token and 45% increase in output speed, according to the Artificial Analysis benchmark while maintaining similar or better quality. This low latency is needed for high-frequency workflows, making it an ideal model for developers to build responsive, real-time experiences.
3.1 Flash-Lite achieves an impressive Elo score of 1432 on the Arena.ai Leaderboard and outperforms other models of similar tier across reasoning and multimodal understanding benchmarks, including 86.9% on GPQA Diamond and 76.8% on MMMU Pro–even surpassing larger Gemini models from prior generations like 2.5 Flash.
Adaptive intelligence at scale for developers
Beyond its raw performance, Gemini 3.1 Flash-Lite comes standard with thinking levels in AI Studio and Vertex AI, giving developers the control and flexibility to select how much the model “thinks” for a task, which is critical for managing high-frequency workloads. 3.1 Flash-Lite can tackle tasks at scale, like high-volume translation and content moderation, where cost is a priority. And it can also handle more complex workloads where more in-depth reasoning is needed, like generating user interfaces and dashboards, creating simulations or following instructions.
Early-access developers on AI Studio and Vertex AI, and companies like Latitude, Cartwheel and Whering are already using 3.1 Flash-Lite to solve complex problems at scale. Early testers highlighted 3.1 Flash-Lite’s efficiency and reasoning capabilities, saying it can handle complex inputs with the precision of a larger-tier model, plus follow instructions and maintain adherence.
We look forward to seeing what you build with 3.1 Flash-Lite and the rest of the Gemini 3 series models.
Original source Report a problem - Feb 26, 2026
- Date parsed from source:Feb 26, 2026
- First seen by Releasebot:Apr 4, 2026
Nano Banana 2: Combining Pro capabilities with lightning-fast speed
Gemini launches Nano Banana 2, a new image generation model that brings advanced world knowledge, stronger instruction following, subject consistency and production-ready specs at Flash speed. It’s rolling out across Gemini, Search, AI Studio, API, Cloud, Flow and Google Ads.
Our latest image generation model offers advanced world knowledge, production-ready specs, subject consistency and more, all at Flash speed.
In August of last year, our Gemini Image model, Nano Banana, became a viral sensation, redefining image generation and editing. Then in November, we released Nano Banana Pro, offering users advanced intelligence and studio-quality creative control. Today, we’re bringing the best of both worlds to users across Google.
Introducing Nano Banana 2 (Gemini 3.1 Flash Image), our latest state-of-the-art image model. Now you can get the advanced world knowledge, quality and reasoning you love in Nano Banana Pro, at lightning-fast speed.
Intelligence and visual quality at Flash speed
Nano Banana 2 brings the high-speed intelligence of Gemini Flash to visual generation, making rapid edits and iteration possible. It makes once-exclusive Pro features accessible to a wider audience, including:
- Advanced world knowledge: The model pulls from Gemini’s real-world knowledge base, and is powered by real-time information and images from web search to more accurately render specific subjects. This deep understanding also helps you create infographics, turn notes into diagrams and generate data visualizations.
- Precision text rendering and translation: Nano Banana 2 allows you to generate accurate, legible text for marketing mockups or greeting cards. You can even translate and localize text within an image to share your ideas globally.
Enhanced creative control
Nano Banana 2 also dramatically closes the gap between speed and visual fidelity, delivering high-quality, photorealistic imagery. Here’s what our newest model offers and has improved on from the original Nano Banana:
- Subject consistency: Maintain character resemblance of up to five characters and the fidelity of up to 14 objects in a single workflow, allowing you to storyboard and build narratives without altering the appearance of your inputs.
- Precise instruction following: With enhanced instruction following, the model adheres more strictly to your complex requests, capturing the specific nuances of your idea so the image you get is the image you asked for.
- Production-ready specs: Make attention grabbing assets with full control of various aspect ratios and resolutions from 512px to 4K, ensuring your visuals stay sharp whether they are for a vertical social post or a wide-screen backdrop.
- Visual fidelity upgrade: Nano Banana 2 delivers vibrant lighting, richer textures and sharper details, maintaining high-quality aesthetics at the speed expected from Flash.
Try Nano Banana 2 today
Whatever your needs, we now offer the perfect tool for every workflow: Nano Banana Pro for high-fidelity tasks requiring maximum factual accuracy, or Nano Banana 2 for rapid generation, precise instruction following and integrated image-search grounding.
Nano Banana 2 is rolling out today across Google products, including:
- Gemini app: Nano Banana 2 will replace Nano Banana Pro across the Fast, Thinking and Pro models. Google AI Pro and Ultra subscribers will keep access to Nano Banana Pro for specialized tasks by regenerating images via the three-dot menu.
- Search: In AI Mode and Lens, through the Google app as well as mobile and desktop browsers. View availability here, including 141 new countries and territories and eight additional languages.
- AI Studio + API: Available in preview in AI Studio and Gemini API. Pricing here. Also available in Google Antigravity.
- Google Cloud: Available in preview with the Gemini API in Vertex AI.
- Flow: Nano Banana 2 is the new default image generation model in Flow, available to all Flow users for zero credits.
- Google Ads: Nano Banana 2 is available now, powering suggestions while creating campaigns in Google Ads.
Robust provenance: marking and verification
As generative media evolves, so must the tools we use to identify and understand it. We continue to deepen our provenance approach, by coupling our state-of-the-art SynthID technology with interoperable C2PA Content Credentials, we provide users with a more holistic and contextual view of not just if AI was used, but how.
Our provenance tools are already making an impact. Since its launch in November, our SynthID verification feature in Gemini app has been used over 20 million times across various languages, helping people identify Google AI-generated images, video and audio. We’ll soon be bringing C2PA verification to the Gemini app, too.
Original source Report a problem - Feb 19, 2026
- Date parsed from source:Feb 19, 2026
- First seen by Releasebot:Apr 4, 2026
Gemini 3.1 Pro: A smarter model for your most complex tasks
Gemini releases Gemini 3.1 Pro, bringing upgraded core intelligence and stronger reasoning across consumer, developer, and enterprise products. It is now rolling out in preview through the Gemini app, NotebookLM, Gemini API, AI Studio, Vertex AI, Gemini Enterprise, Gemini CLI, Antigravity, and Android Studio.
Last week, we released a major update to Gemini 3 Deep Think to solve modern challenges across science, research and engineering. Today, we’re releasing the upgraded core intelligence that makes those breakthroughs possible: Gemini 3.1 Pro. We are shipping 3.1 Pro across our consumer and developer products to bring this progress in intelligence to your everyday applications.
Starting today, 3.1 Pro is rolling out:
- For developers in preview via the Gemini API in Google AI Studio, Gemini CLI, our agentic development platform Google Antigravity and Android Studio
- For enterprises in Vertex AI and Gemini Enterprise
- For consumers via the Gemini app and NotebookLM
Building on the Gemini 3 series, 3.1 Pro represents a step forward in core reasoning. 3.1 Pro is a smarter, more capable baseline for complex problem-solving. This is reflected in our progress on rigorous benchmarks. On ARC-AGI-2, a benchmark that evaluates a model’s ability to solve entirely new logic patterns, 3.1 Pro achieved a verified score of 77.1%. This is more than double the reasoning performance of 3 Pro.
3.1 Pro is designed for tasks where a simple answer isn’t enough, taking advanced reasoning and making it useful for your hardest challenges. This improved intelligence can help in practical applications — whether you’re looking for a clear, visual explanation of a complex topic, a way to synthesize data into a single view, or bringing a creative project to life.
Since releasing Gemini 3 Pro in November, your feedback and the pace of progress have driven these rapid improvements. We are releasing 3.1 Pro in preview today to validate these updates and continue to make further advancements in areas such as ambitious agentic workflows before we make it generally available soon.
Starting today, Gemini 3.1 Pro in the Gemini app is rolling out with higher limits for users with the Google AI Pro and Ultra plans. 3.1 Pro is also now available on NotebookLM exclusively for Pro and Ultra users. And developers and enterprises can access 3.1 Pro now in preview in the Gemini API via AI Studio, Antigravity, Vertex AI, Gemini Enterprise, Gemini CLI and Android Studio.
We can’t wait to see what you build and discover with it.
Original source Report a problem - Feb 18, 2026
- Date parsed from source:Feb 18, 2026
- First seen by Releasebot:Apr 4, 2026
A new way to express yourself: Gemini can now create music
Gemini adds Lyria 3 music generation in beta, letting users create 30-second custom tracks from text or images, with richer style control, auto-generated lyrics, and SynthID audio verification. It’s rolling out on desktop and mobile, with higher limits for AI Plus, Pro and Ultra subscribers.
The Gemini app now features our most advanced music generation model Lyria 3, empowering anyone to make 30-second tracks using text or images in beta.
Since launching the Gemini app, we've built tools to encourage creative expression through images and video. Today, we're taking the next step: custom music generation. Lyria 3, Google DeepMind’s latest generative music model, is rolling out today in beta in the Gemini app. Just describe an idea or upload a photo, like “a comical R&B slow jam about a sock finding their match" and in a matter of seconds, Gemini will translate it into a high-quality, catchy track. To push the creative envelope further, you can even ask Gemini to take inspiration from something you upload.
Lyria 3 improves on audio generation from our Lyria models in three important ways:
- No need to provide your own lyrics! They'll be generated for you based on your prompt.
- You have more creative control over elements like the style, vocals and tempo you want.
- You can create more realistic and musically complex tracks.
Here’s how you can use it:
- Text to track: Describe a specific genre, mood, inside joke, or memory to create unique tracks with lyrics or instrumental audio that fits your vibe. “I’m feeling nostalgic. Create a track for my mother about the great times we had as kids and the memories of her home cooked plantains. Make it a fun afrobeat track with a true African vibe.”
- From photos and videos to track: Upload a photo or video and watch Gemini use the content to compose a track with lyrics that fit the mood perfectly. “Use these photos to create a track about my dog Duncan on a hike in the woods.”
The Gemini app creates 30-second tracks with custom cover art generated by Nano Banana. This makes it easy to quickly share with friends by downloading or simply clicking the share link. The goal of these tracks isn't to create a musical masterpiece, but rather to give you a fun, unique way to express yourself.
Creators can also explore Lyria 3 on YouTube’s Dream Track. Available in the U.S. and now rolling out to YouTube creators in other countries, Lyria 3 will enhance the quality of each unique Shorts soundtrack. Whether it's creating a lyrical verse or a vibey backing track, being able to better customize the soundtrack will take creators’ Shorts to the next level.
New audio verification capabilities
All tracks generated in the Gemini app are embedded with SynthID, our imperceptible watermark for identifying Google AI-generated content. We are also giving you more tools to help identify AI content, broadening our verification capabilities in the Gemini app to include audio, along with image and video. Simply upload a file and ask if it was generated using Google AI, and Gemini will check for SynthID and use its own reasoning to return a response.
Our commitment to developing generative AI responsibly
Since we first launched Lyria in 2023, we've sought to develop this technology responsibly in collaboration with the music community. We've learned a lot through these collaborations and our experiments, like Music AI Sandbox, and have been very mindful of copyright and partner agreements as we've trained Lyria 3.
Music generation with Lyria 3 is designed for original expression, not for mimicking existing artists. If your prompt names a specific artist, Gemini will take this as broad creative inspiration and create a track that shares a similar style or mood. We also have filters in place to check outputs against existing content. We recognize that our approach might not be foolproof, so you can report content that may violate your rights or the rights of others. Additionally, in order to use our products, users must adhere to our Terms of Service and Gen AI prohibited use policies, which prohibit violations of others’ intellectual property and privacy rights.
Lyria 3 is available in the Gemini app for all users 18+ in English, German, Spanish, French, Hindi, Japanese, Korean and Portuguese, with plans to expand quality and coverage of more languages, rolling out on desktop today and to the mobile app over the next several days. And Google AI Plus, Pro and Ultra subscribers will enjoy higher limits.
Our goal with music generation in the Gemini app is to help you add a fun, custom soundtrack to your daily life. Try it out today at gemini.google.com.
Original source Report a problem - Jan 13, 2026
- Date parsed from source:Jan 13, 2026
- First seen by Releasebot:Apr 4, 2026
Veo 3.1 Ingredients to Video: More consistency, creativity and control
Gemini adds Veo 3.1 video upgrades with more expressive Ingredients to Video, native vertical outputs, and upscaling to 1080p and 4K, plus expanded access across the Gemini app, YouTube, Flow, Google Vids, the Gemini API, and Vertex AI. It also adds video verification in the app.
Our latest Veo update generates lively, dynamic clips that feel natural and engaging — and supports vertical video generation.
Today, Veo is getting more expressive, with improvements that help you create more fun, creative, high-quality videos based on ingredient images, built directly for the mobile format. We’re excited to bring new creative possibilities for everyone from casual storytellers to professional filmmakers.
We’re releasing:
- Improvements to Veo 3.1 Ingredients to Video, our capability that lets you create videos based on reference images. This update makes videos more expressive and creative, even with simple prompts
- Native vertical outputs for Ingredients to Video (portrait mode) to power mobile-first, short-form video creation
- State-of-the-art upscaling to 1080p and 4K resolution for high-fidelity production workflows
Whether you are looking for livelier movement, better control over visual elements or broadcast-ready resolution, these updates give you the tools to bring your vision to life. These updates are launching in the Gemini app, YouTube, Flow, Google Vids, the Gemini API and Vertex AI.
Improvements to Veo 3.1 Ingredients to Video
Turn ingredient images into fun, shareable clips
Even with short prompts, you can generate dynamic and engaging videos based on ingredient images. You’ll now see richer dialogue and storytelling, making your videos feel more alive and expressive.
Maintain identity consistency for your characters
Identity consistency is better than ever with Veo 3.1 Ingredients to Video. Keep your characters looking the same even as the setting changes, making it easier to tell a full narrative by having the same character appear across multiple scenes.
Achieve background and object consistency
Control the scene by maintaining the integrity of your setting and the objects within it. You can also reuse an object, backgrounds or textures across scenes.
Seamlessly blend textures, characters and objects
Combine disparate elements — like characters, objects, textures and stylized backgrounds — into a cohesive, high-impact clip.
Pro tip: use the new Nano Banana Pro (Gemini 3 Pro Image) in the Gemini app or Flow to create your ingredient images, which you can then use to create stunning videos with Veo 3.1 Ingredients to Video.
Create high-fidelity visuals with upgraded capabilities
With Veo 3.1’s new capabilities, we are introducing mobile-optimized outputs and professional-grade quality options.
Native vertical outputs for Ingredients to Video
For the first time, "Ingredients to Video" supports generating videos in a native 9:16 aspect ratio. Whether you are creating for YouTube Shorts or other platforms, you can now produce high-quality, full-screen vertical storytelling without cropping or quality loss.
State-of-the-art upscaling to 1080p and 4K resolution
Generate videos 1080p and 4K with state-of-the-art upscaling. Our improved 1080p resolution offers a sharper, cleaner video perfect for editing. For even more detail, choose 4K to capture rich textures and stunning clarity — ideal for high-end productions and large screens.
Try these updates today
Across our products and services, you can now access these new capabilities tailored to your workflow:
- Consumers and creators: We are bringing Veo 3.1 Ingredients to Video directly to YouTube Shorts and the YouTube Create app for the first time. You can also try the enhanced Veo 3.1 Ingredients to Video and portrait mode for Veo in the Gemini app starting today.
- Professional and enterprise workflows: The enhanced Veo 3.1 Ingredients to Video and native vertical format support are rolling out to Flow, the Gemini API, Vertex AI, and Google Vids, with 1080p and 4K resolution options also available on Flow, the API, and Vertex AI.
Verify videos in the Gemini app
We’re committed to providing tools to make it easier to determine if content is AI-generated. This is why videos generated by Google’s tools are embedded with our imperceptible SynthID digital watermark.
In December we expanded our powerful verification tool in the Gemini app to include video. You can now upload a video and simply ask if it was generated with Google AI. This builds on our existing image verification tools, helping to foster a more transparent ecosystem for everyone.
You can find out more about how we’re increasing transparency in AI content with SynthID in our blog post.
Original source Report a problem - Dec 19, 2025
- Date parsed from source:Dec 19, 2025
- First seen by Releasebot:Apr 4, 2026
Gemma Scope 2: helping the AI safety community deepen understanding of complex language model behavior
Gemini releases Gemma Scope 2, an open suite of interpretability tools for Gemma 3 models from 270M to 27B parameters, giving researchers deeper visibility into model behavior, safety risks, and chatbot reasoning.
Announcing a new, open suite of tools for language model interpretability
Large Language Models (LLMs) are capable of incredible feats of reasoning, yet their internal decision-making processes remain largely opaque. Should a system not behave as expected, a lack of visibility into its internal workings can make it difficult to pinpoint the exact reason for its behaviour. Last year, we advanced the science of interpretability with Gemma Scope, a toolkit designed to help researchers understand the inner workings of Gemma 2, our lightweight collection of open models.
Today, we are releasing Gemma Scope 2: a comprehensive, open suite of interpretability tools for all Gemma 3 model sizes, from 270M to 27B parameters. These tools can enable us to trace potential risks across the entire "brain" of the model.
To our knowledge, this is the largest ever open-source release of interpretability tools by an AI lab to date. Producing Gemma Scope 2 involved storing approximately 110 Petabytes of data, as well as training over 1 trillion total parameters.
As AI continues to advance, we look forward to the AI research community using Gemma Scope 2 to debug emergent model behaviors, use these tools to better audit and debug AI agents, and ultimately, accelerate the development of practical and robust safety interventions against issues like jailbreaks, hallucinations and sycophancy.
Our interactive Gemma Scope 2 demo is available to try, courtesy of Neuronpedia.
What’s new in Gemma Scope 2
Interpretability research aims to understand the internal workings and learned algorithms of AI models. As AI becomes increasingly more capable and complex, interpretability is crucial for building AI that is safe and reliable.
Like its predecessor, Gemma Scope 2 acts as a microscope for the Gemma family of language models. By combining sparse autoencoders (SAEs) and transcoders, it allows researchers to look inside models, see what they’re thinking about, and how these thoughts are formed and connect to the model’s behaviour. In turn, this enables the richer study of jailbreaks or other AI behaviours relevant to safety, like discrepancies between a model's communicated reasoning and its internal state.
While the original Gemma Scope enabled research in key areas of safety, such as model hallucination, identifying secrets known by a model, and training safer models, Gemma Scope 2 supports even more ambitious research through significant upgrades:
- Full coverage at scale: We provide a full suite of tools for the entire Gemma 3 family (up to 27B parameters), essential for studying emergent behaviors that only appear at scale, such as those previously uncovered by the 27b-size C2S Scale model that helped discover a new potential cancer therapy pathway. Although Gemma Scope 2 is not trained on this model, this is an example of the kind of emergent behavior that these tools might be able to understand.
- More refined tools to decipher complex internal behaviors: Gemma Scope 2 includes SAEs and transcoders trained on every layer of our Gemma 3 family of models. Skip-transcoders and Cross-layer transcoders make it easier to decipher multi-step computations and algorithms spread throughout the model.
- Advanced training techniques: We use state-of-the-art techniques, notably the Matryoshka training technique, which helps SAEs detect more useful concepts and resolves certain flaws discovered in Gemma Scope.
- Chatbot behavior analysis tools: We also provide interpretability tools targeted at the versions of Gemma 3 tuned for chat use cases. These tools enable analysis of complex, multi-step behaviors, such as jailbreaks, refusal mechanisms, and chain-of-thought faithfulness.
Advancing the field
By releasing Gemma Scope 2, we aim to enable the AI safety research community to push the field forward using a suite of cutting-edge interpretability tools. This new level of access is crucial for tackling real-world safety problems that only arise in larger, modern LLMs.
Learn more about Gemma Scope
Original source Report a problem
Download Gemma Scope 2
View our models on Neuronpedia
Read our technical report
Try our Colab tutorial
View our Gemma Scope page - Dec 17, 2025
- Date parsed from source:Dec 17, 2025
- First seen by Releasebot:Apr 4, 2026
Gemini 3 Flash: frontier intelligence built for speed
Gemini expands its model family with Gemini 3 Flash, a fast, lower-cost model that brings Pro-grade reasoning to the Gemini app, AI Mode in Search, Google AI Studio, Gemini CLI, Google Antigravity, Vertex AI and Gemini Enterprise.
Gemini 3 Flash is our latest model with frontier intelligence built for speed that helps everyone learn, build, and plan anything — faster.
Today, we're expanding the Gemini 3 model family with the release of Gemini 3 Flash, which offers frontier intelligence built for speed at a fraction of the cost. With this release, we’re making Gemini 3’s next-generation intelligence accessible to everyone across Google products.
Last month, we kicked off Gemini 3 with Gemini 3 Pro and Gemini 3 Deep Think mode, and the response has been incredible. Since launch day, we have been processing over 1T tokens per day on our API. We’ve seen you use Gemini 3 to vibe code simulations to learn about complex topics, build and design interactive games and understand all types of multimodal content.
With Gemini 3, we introduced frontier performance across complex reasoning, multimodal and vision understanding, and agentic and vibe coding tasks. Gemini 3 Flash retains this foundation, combining Gemini 3's Pro-grade reasoning with Flash-level latency, efficiency and cost. It not only enables everyday tasks with improved reasoning, but also is our most impressive model for agentic workflows.
Starting today, Gemini 3 Flash is rolling out to millions of people globally:
- For developers in the Gemini API in Google AI Studio, Gemini CLI and our new agentic development platform Google Antigravity
- For everyone via the Gemini app and in AI Mode in Search
- For enterprises in Vertex AI and Gemini Enterprise
Gemini 3 Flash: frontier intelligence at scale
Gemini 3 Flash demonstrates that speed and scale don’t have to come at the cost of intelligence. It delivers frontier performance on PhD-level reasoning and knowledge benchmarks like GPQA Diamond (90.4%) and Humanity’s Last Exam (33.7% without tools), rivaling larger frontier models, and significantly outperforming even the best 2.5 model, Gemini 2.5 Pro, across a number of benchmarks. It also reaches state-of-the-art performance with an impressive score of 81.2% on MMMU Pro, comparable to Gemini 3 Pro.
In addition to its frontier-level reasoning and multimodal capabilities, Gemini 3 Flash was built to be highly efficient, pushing the Pareto frontier of quality vs. cost and speed. When processing at the highest thinking level, Gemini 3 Flash is able to modulate how much it thinks. It may think longer for more complex use cases, but it also uses 30% fewer tokens on average than 2.5 Pro, as measured on typical traffic, to accurately complete everyday tasks with higher performance.
Gemini 3 Flash pushes the Pareto frontier on performance vs. cost and speed. Performance, here, is measured by LMArena Elo Score.
Gemini 3 Flash’s strength lies in its raw speed, building on the Flash series that developers and consumers already love. It outperforms 2.5 Pro while being 3x faster (based on Artificial Analysis benchmarking) at a fraction of the cost. Gemini 3 Flash is priced at $0.50/1M input tokens and $3/1M output tokens (audio input remains at $1/1M input tokens).
For developers: intelligence that keeps up
Gemini 3 Flash is made for iterative development, offering Gemini 3’s Pro-grade coding performance with low latency — it’s able to reason and solve tasks quickly in high-frequency workflows. On SWE-bench Verified, a benchmark for evaluating coding agent capabilities, Gemini 3 Flash achieves a score of 78%, outperforming not only the 2.5 series, but also Gemini 3 Pro. It strikes an ideal balance for agentic coding, production-ready systems and responsive interactive applications.
Gemini 3 Flash in Google Antigravity works quickly to update production-ready applications.
Gemini 3 Flash’s strong performance in reasoning, tool use and multimodal capabilities is ideal for developers looking to do more complex video analysis, data extraction and visual Q&A, which means it can enable more intelligent applications — like in-game assistants or A/B test experiments — that demand both quick answers and deep reasoning.
We’ve received a tremendous response from companies using Gemini 3 Flash. Companies like JetBrains, Bridgewater Associates, and Figma are already using it to transform their businesses, recognizing how its inference speed, efficiency and reasoning capabilities perform on par with larger models. Gemini 3 Flash is available today to enterprises via Vertex AI and Gemini Enterprise.
For everyone: Gemini 3 Flash is rolling out globally
Gemini 3 Flash is now the default model in the Gemini app, replacing 2.5 Flash. That means all of our Gemini users globally will get access to the Gemini 3 experience at no cost, giving their everyday tasks a major upgrade.
Because of Gemini 3 Flash’s incredible multimodal reasoning capabilities, you can use it to help you see, hear and understand any type of information faster. For example, you can ask Gemini to understand your videos and images and turn that content into a helpful and actionable plan in just a few seconds.
Or you can quickly build fun, useful apps from scratch using your voice without prior coding knowledge. Just dictate to Gemini on the go, and it can transform your unstructured thoughts into a functioning app in minutes.
Gemini 3 Flash is also starting to roll out as the default model for AI Mode in Search with access to everyone around the world.
Building on the reasoning capabilities of Gemini 3 Pro, AI Mode with Gemini 3 Flash is more powerful at parsing the nuances of your question. It considers each aspect of your query to serve thoughtful, comprehensive responses that are visually digestible — pulling real-time local information and helpful links from across the web. The result effectively combines research with immediate action: you get an intelligently organized breakdown alongside specific recommendations — at the speed of Search.
This shines when tackling complex goals with multiple considerations like trying to plan a last-minute trip or learning complex educational concepts quickly.
Gemini 3 Flash brings the incredible reasoning capabilities of Gemini 3 to Search, without compromising speed, so you can tackle your most complicated questions.
Try Gemini 3 Flash today
Gemini 3 Flash is available now in preview via the Gemini API in Google AI Studio, Google Antigravity, Vertex AI and Gemini Enterprise. You can also access it through other developer tools like Gemini CLI and Android Studio. It’s also starting to roll out to everyone in the Gemini app and AI Mode in Search, bringing fast access to next-generation intelligence at no cost.
We’re looking forward to seeing what you bring to life with this expanded family of models: Gemini 3 Pro, Gemini 3 Deep Think and now, Gemini 3 Flash.
Original source Report a problem
This is the end. You've seen all the release notes in this feed!