Google DeepMind Enhances Veo with 3.1 Update and Flow Upgrades

New features offer creators more control, richer audio, and enhanced realism in AI-generated video.

Google DeepMind has rolled out significant updates to its Veo video generation platform, introducing Veo 3.1 and enhanced capabilities in Flow. These improvements bring richer audio, greater narrative control, and advanced editing tools for creators.

Katie Rowan

By Katie Rowan

December 5, 2025

4 min read

Google DeepMind Enhances Veo with 3.1 Update and Flow Upgrades

Key Facts

  • Veo 3.1 introduces richer audio, more narrative control, and enhanced realism.
  • Flow now supports audio across all features and offers advanced editing capabilities.
  • New features include 'Ingredients to Video' for style control and 'Frames to Video' for seamless transitions.
  • The 'Extend' feature allows for creating longer, seamless video shots, up to a minute or more.
  • Flow can now add new elements to scenes with natural lighting and will soon remove unwanted objects seamlessly.

Why You Care

Ever wished you could sculpt your video ideas with the precision of a master artist, even if you’re using AI? Google DeepMind’s latest updates to Veo and Flow are here to make that a reality. What if you could tell your story exactly as you envision it, with AI handling the complex details? These advancements promise to give you creative control over your AI-generated videos.

What Actually Happened

Google DeepMind has announced significant updates to its video generation system, Veo, and its companion editing environment, Flow. The company reports that Flow now features enhanced creative tools and supports audio across all functionalities. What’s more, Veo 3.1 introduces richer audio, more narrative control, and enhanced realism, capturing true-to-life textures, as mentioned in the release. These updates aim to provide users with more granular control over their final video scenes. For the first time, audio capabilities are extending to existing features like “Ingredients to Video,” “Frames to Video,” and “Extend,” according to the announcement.

Why This Matters to You

These updates mean you can now refine your AI-generated videos with much greater ease and accuracy. Imagine crafting a scene where every character, object, and style element perfectly matches your vision. You can now use multiple reference images to achieve this with “Ingredients to Video,” the team revealed. This allows Flow to create a final scene that looks just as you envisioned, as detailed in the blog post.

What kind of visual stories will you be able to create with this new level of precision?

“We’re always listening to your feedback, and we’ve heard that you want more artistic control within Flow, with increased support for audio across all features,” states Jess Gallegos, Senior Product Manager at Google DeepMind. This feedback directly led to these new tools.

Here’s a quick look at some key new capabilities:

  • Craft the look of your scene: Use multiple reference images to control characters, objects, and style.
  • Control the shot from start to finish: Provide a starting and ending image to generate video transitions.
  • Create longer, shots: Extend videos for a minute or more, connecting and continuing the action.
  • Add new elements to any scene: Introduce realistic details or fantastical creatures with natural lighting.

The Surprising Finding

One particularly interesting creation is the object manipulation within Flow. The documentation indicates that Flow now handles complex details like shadows and scene lighting when adding new elements. This makes additions look natural. This is quite surprising because integrating new elements seamlessly into existing AI-generated content is a significant technical challenge. Often, such additions can look artificial or out of place. The ability to reconstruct backgrounds when removing objects, making it appear as though the object was never there, is also a remarkable leap. This challenges the common assumption that AI-generated video editing would always leave noticeable artifacts. It suggests a much deeper understanding of scene composition by the AI.

What Happens Next

These experimental features are actively improving, and Google DeepMind is keen to iterate based on user feedback. You can expect to see these capabilities refined over the coming months. For example, the ability to remove unwanted objects or characters seamlessly is “coming soon.” This suggests a rollout within the next few quarters. Imagine being able to film a scene and then effortlessly remove an accidental photobomber. This will be a huge boon for content creators. The company reports that each extended video is generated based on the final second of your previous clip. This makes it most useful for creating a longer establishing shot.

Content creators, filmmakers, and marketers should start exploring these tools now. Understanding how to integrate these precise controls into your workflow will be crucial. This will allow you to stay ahead in the rapidly evolving world of AI-driven video production. The industry implications are clear: higher quality, more personalized, and more easily editable AI-generated video content is becoming the new standard.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice