Google DeepMind Unveils Gemini Robotics: A Leap Toward More Adaptive and Interactive Robots
Google DeepMind is taking AI beyond screens into the real world with Gemini Robotics, a new model built on Gemini 2.0 that allows robots to see, understand, and interact with their surroundings. This update introduces two AI-driven models, Gemini Robotics and Gemini Robotics-ER which are designed to improve robots' ability to perceive, interact with, and respond intelligently to their surroundings.
For any robots to be truly useful, they need more than just computation; they must be able to see, understand, and act in dynamic environments. Gemini Robotics is built on a vision-language-action (VLA) framework, enabling it to interpret the world, respond to commands, and take physical action.
Key Features:
Gemini Robotics-ER: Enhancing Spatial Intelligence
Alongside Gemini Robotics, DeepMind is introducing Gemini Robotics-ER, a model designed to improve spatial understanding and reasoning in robotics. This extension allows robots to:
By bridging high-level AI reasoning with low-level robot control systems, Gemini Robotics-ER enables roboticists to integrate its capabilities into existing robotic frameworks more efficiently.
Advancing AI Safety in Robotics
As AI-driven robots become more capable, safety remains a priority. DeepMind is implementing multiple layers of protection:
DeepMind is also introducing ASIMOV, a dataset designed to evaluate and improve AI safety in robotic applications, reinforcing a structured approach to responsible AI development.
The future of robotics isn’t just about automation—it’s about AI that perceives, learns, and collaborates with us in everyday life.
About the Author
Mia Cruz
Mia Cruz is an AI news correspondent from United States of America.
Recent Articles
Subscribe to Newsletter
Enter your email address to register to our newsletter subscription!