Runtime Learning of Quadruped Robots in Wild Environments

Cai, Yihao, Mao, Yanbing, Sha, Lui, Cao, Hongpeng, Caccamo, Marco

arXiv.org Artificial Intelligence 

Runtime Learning of Quadruped Robots in Wild Environments Yihao Cai 1, Y anbing Mao 1, Lui Sha 2, Hongpeng Cao 3, and Marco Caccamo 3 This paper presents a runtime learning framework for quadruped robots, enabling them to learn and adapt safely in dynamic wild environments. The core novelty of this framework lies in two interactive and complementary components within the control module: the high-performance (HP)-Student and the high-assurance (HA)-Teacher. HP-Student is a deep reinforcement learning (DRL) agent that engages in self-learning and teaching-to-learn to develop a safe and high-performance action policy. HA-Teacher is a simplified yet verifiable physics-model-based controller, with the role of teaching HP-Student about safety while providing a backup for the robot's safe locomotion. HA-Teacher is innovative due to its real-time physics model, real-time action policy, and real-time control goals, all tailored to respond effectively to real-time wild environments, ensuring safety. The framework also includes a coordinator who effectively manages the interaction between HP-Student and HA-Teacher. Experiments involving a Unitree Go2 robot in Nvidia Isaac Gym and comparisons with state-of-the-art safe DRLs demonstrate the effectiveness of the proposed runtime learning framework. I NTRODUCTION Quadruped robots have become a promising solution for navigating challenging wild environments, such as forests, disaster zones, and mountainous regions [1], [2].