Gemini Robotics: AI Models Step into the Physical World

Google DeepMind introduces new Gemini 2.0-based models designed to power advanced robotics.

Google DeepMind has unveiled Gemini Robotics and Gemini Robotics-ER, two new AI models built on Gemini 2.0. These models aim to bring advanced AI capabilities into the physical world, enabling robots to perform complex tasks with human-like understanding and dexterity. This marks a significant step towards more helpful and general-purpose robots.

Mark Ellison

By Mark Ellison

December 5, 2025

4 min read

Gemini Robotics: AI Models Step into the Physical World

Key Facts

  • Google DeepMind introduced two new AI models: Gemini Robotics and Gemini Robotics-ER.
  • These models are built on the Gemini 2.0 architecture and are designed for physical robotics.
  • Gemini Robotics is a vision-language-action (VLA) model that directly controls robots.
  • Gemini Robotics-ER offers advanced spatial understanding for embodied reasoning.
  • Google DeepMind is partnering with Apptronik to build humanoid robots with Gemini 2.0.

Why You Care

Ever wondered when robots would truly become helpful companions in our daily lives? Google DeepMind just took a giant leap forward. They’ve introduced Gemini Robotics, a new collection of AI models designed to bring intelligence into the physical world. This isn’t just about digital smarts anymore. It’s about robots that can see, understand, and act in your environment. Why should this matter to you? Because these advancements could soon impact everything from manufacturing to personal assistance.

What Actually Happened

Google DeepMind has officially launched two new AI models based on their Gemini 2.0 architecture, as mentioned in the release. These models are specifically engineered for robotics. The first, Gemini Robotics, is an vision-language-action (VLA) model. It uses physical actions as a new output modality, allowing it to directly control robots. The second model is Gemini Robotics-ER, which stands for Embodied Reasoning. This model provides spatial understanding, empowering roboticists to develop programs using Gemini’s embodied reasoning capabilities. These innovations move AI beyond the digital realm and into tangible, real-world applications, according to the announcement.

Why This Matters to You

These new Gemini Robotics models promise to make robots far more capable and adaptable. Imagine a robot that can understand complex instructions and perform intricate tasks, not just repetitive motions. The company reports that these models enable a wider range of real-world tasks than previously possible. For example, think of a robot in a warehouse. Instead of needing precise programming for every item, it could learn to identify and sort new objects on the fly.

Google DeepMind is also collaborating with Apptronik to develop the next generation of humanoid robots using Gemini 2.0. What’s more, they are working with trusted testers to refine Gemini Robotics-ER. This collaboration suggests practical applications are on the horizon. What kind of everyday tasks could you envision a more intelligent, dexterous robot helping you with?

Carolina Parada stated, “In order for AI to be useful and helpful to people in the physical realm, they have to demonstrate ‘embodied’ reasoning — the humanlike ability to comprehend and react to the world around us— as well as safely take action to get things done.” This highlights the core goal: making robots truly intuitive and helpful.

Here’s a quick look at the key qualities these new models aim to deliver:

QualityDescription
GeneralityAdapts to diverse situations, even those not seen in training.
InteractivityUnderstands and responds quickly to instructions and environmental changes.
DexterityPerforms precise manipulations with hands and fingers, like humans.

The Surprising Finding

Perhaps the most compelling aspect of Gemini Robotics is its ability to generalize to novel situations. The technical report explains that Gemini Robotics can solve a wide variety of tasks “out of the box,” including those it has never seen before in training. This challenges the common assumption that robots need extensive, specific training for every single new task or object. For instance, a robot trained to pick up blocks might now, surprisingly, be able to manipulate a screwdriver it has never encountered. The research shows this includes dealing with new objects, diverse instructions, and unfamiliar environments. This level of adaptability is a significant leap, pushing us closer to truly general-purpose robots that don’t require constant re-programming for minor variations.

What Happens Next

The future will involve continued exploration and creation of these models’ capabilities. Google DeepMind is actively partnering with companies like Apptronik and engaging trusted testers, suggesting a strong focus on real-world applications. We can expect to see more demonstrations and pilot programs within the next 12 to 18 months. For example, imagine these robots assisting in complex assembly lines or even in disaster relief scenarios, adapting to unpredictable environments. For you, this means a future where robots are less like rigid machines and more like flexible assistants.

Industry implications are vast, impacting manufacturing, logistics, and even personal services. Companies should start considering how adaptable robots could integrate into their operations. The team revealed they are “looking forward to exploring our models’ capabilities and continuing to develop them on the path to real-world applications.” This indicates a continuous push towards practical, impactful robotic solutions, powered by Gemini Robotics AI.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice