Google Deep Mind on Tuesday released a new Gemini robotics artificial intelligence (AI) model that can completely run on a local device. Dodged Gemini Robotics on Device, it is a voice -language action (VLA) model that can carry the robot widely in the real world environment. Since the AI model works without the need to be associated with the data network, it is more useful to the applications that are delayed, said Mountain View -based Tech Dev. Currently, the model is available to those who have signed up for its reliable tester program.
Google’s new robotics model runs fully on a device
In a blog post, Carolina Parada announced the release of the Gemini Robotics on Device, senior director of Google Deep Mind and head of robotics. The new VLA model can be obtained through the Gemini Robotics Software Development Kit (SDK) after signing up to its tester program. The model can also be tested on the company’s Majoko Physics Simulator.
Since it is a proprietary model, details about its architecture and training methods are not known. However, Google has highlighted its abilities. The VLA model is designed for two arm robots and has the least computational requirements. Nevertheless, the model allows experiences, and the company claims that it can adapt to new tasks with only 50 to 100 demonstrations.
The Gemini Robotics on Device also follows natural language guidelines and can perform complex tasks such as these zapping bags or folding fabrics. Based on internal testing, Tech giant claims that the AI model “shows stronger normal performance locally.” In addition, it is also said to improve other on -device models about “distribution more difficult tasks and complicated multi -faceted instructions.”
Specifically, Google highlighted that when the AI model was trained for the Aluha Robot, researchers also managed to adapt it to Franka FR3 and Apotonic’s Apollo Humanide Robot. These are all two arm robots, the only sequences that are compatible with the Gemini robotics on device.
The AI model was able to follow the instructions on all different robots and perform the general task. The company claimed that it could handle pre -visible items and scenes, and even perform industrial belt assembly works that require high -level precision and expertise.


