GR-2: A Generative Video-Language-Action Model with Web-Scale Knowledge for Robot Manipulation
Cheang, Chi-Lam, Chen, Guangzeng, Jing, Ya, Kong, Tao, Li, Hang, Li, Yifeng, Liu, Yuxiao, Wu, Hongtao, Xu, Jiafeng, Yang, Yichu, Zhang, Hanbo, Zhu, Minzhao
–arXiv.org Artificial Intelligence
GR-2 is first pre-trained on a vast number of Internet videos to capture the dynamics of the world. This large-scale pre-training, involving 38 million video clips and over 50 billion tokens, equips GR-2 with the ability to generalize across a wide range of robotic tasks and environments during subsequent policy learning. Following this, GR-2 is fine-tuned for both video generation and action prediction using robot trajectories. It exhibits impressive multi-task learning capabilities, achieving an average success rate of 97.7% across more than 100 tasks. Moreover, GR-2 demonstrates exceptional generalization to new, previously unseen scenarios, including novel backgrounds, environments, objects, and tasks.
arXiv.org Artificial Intelligence
Oct-8-2024
- Genre:
- Research Report (1.00)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning (1.00)
- Robots > Manipulation (0.51)
- Information Technology > Artificial Intelligence