Google is Working on Gemini Robotics Models, According to Sundar Pichai
Gemini Robotics is an advanced vision-language-action (VLA) model that was built on Gemini 2.0 with the addition of physical actions as a new output modality

During Google Q1 2025 earning calls, Alphabet CEO Sundar Pichai revealed that Google is working on a family of Gemini models for robotics.
"Lastly, we're developing AI models in new areas where there’s enormous opportunity, for example, our new Gemini Robotics models," he said.
Developed by Google Deepmind, its AI research wing, the company is expected to make newer announcements for its new family of Gemini models soon.
Gemini Robotics is an advanced vision-language-action (VLA) model that was built on Gemini 2.0 with the addition of physical actions as a new output modality for the purpose of directly controlling robots.
The company has developed another model called Gemini Robotics-ER, a Gemini model with advanced spatial understanding, enabling roboticists to run their own programs using Gemini’s embodied reasoning (ER) abilities.
"Both of these models enable a variety of robots to perform a wider range of real-world tasks than ever before. As part of our efforts, we’re partnering with Apptronik to build the next generation of humanoid robots with Gemini 2.0. We’re also working with a selected number of trusted testers to guide the future of Gemini Robotics-ER," Google DeepMind said in a blog post.
Last month, Google launched Google’s Gemini 2.5 Pro, which received strong feedback for its advances in reasoning, coding, and math.
Pichai revealed that active usage of AI Studio and the Gemini API has surged 200% this year.
The company also introduced Gemini 2.5 Flash, rolled out new models Imagen 3 and Veo 2, launched Gemma 3, and expanded into robotics and health with AI Co-Scientist and AlphaFold.
Comments ()