Google's Gemini Robotics AI Model Reaches Into the Physical World
Briefly

Google DeepMind unveiled Gemini Robotics, an advanced AI model that combines language comprehension, visual perception, and physical action for greater robot utility. Demonstrations showcased robots adeptly handling tasks such as folding paper and managing objects in response to verbal cues. The new model is designed to generalize behavior across various hardware setups, enhancing adaptability. Additionally, Gemini Robotics-ER focuses on visual and spatial reasoning to support broader robotics research. The initiative marks a significant leap in robotics, allowing robots to perform tasks outside their previous training scenarios seamlessly.
Google DeepMind's new model, Gemini Robotics, integrates language and physical action, enabling robots to understand and execute tasks based on spoken commands.
The Gemini Robotics model seeks to generalize behavior across various hardware, allowing for a unified approach in robotic functions and tasks.
Read at WIRED
[
|
]