GR-2: A Generative Video-Language-Action Model with Web-Scale Knowledge for Robot Manipulation

Cheang, Chi-Lam, Chen, Guangzeng, Jing, Ya, Kong, Tao, Li, Hang, Li, Yifeng, Liu, Yuxiao, Wu, Hongtao, Xu, Jiafeng, Yang, Yichu, Zhang, Hanbo, Zhu, Minzhao

arXiv.org Artificial Intelligence 

GR-2 is first pre-trained on a vast number of Internet videos to capture the dynamics of the world. This large-scale pre-training, involving 38 million video clips and over 50 billion tokens, equips GR-2 with the ability to generalize across a wide range of robotic tasks and environments during subsequent policy learning. Following this, GR-2 is fine-tuned for both video generation and action prediction using robot trajectories. It exhibits impressive multi-task learning capabilities, achieving an average success rate of 97.7% across more than 100 tasks. Moreover, GR-2 demonstrates exceptional generalization to new, previously unseen scenarios, including novel backgrounds, environments, objects, and tasks.