SoftMimic: Learning Compliant Whole-body Control from Examples

Margolis, Gabriel B., Wang, Michelle, Fey, Nolan, Agrawal, Pulkit

arXiv.org Artificial Intelligence 

We train humanoid policies that compliantly respond to external forces while tracking a reference motion. The desired force-displacement relationship is modulated by a'stiffness' input at deployment time, and a single policy learns to realize a wide range of stiffnesses. In the images, the reference motion is visualized in blue, and the approximate external force on the robot is illustrated by the red arrows. Abstract-- We introduce SoftMimic, a framework for learning compliant whole-body control policies for humanoid robots from example motions. Imitating human motions with reinforcement learning allows humanoids to quickly learn new skills, but existing methods incentivize stiff control that aggressively corrects deviations from a reference motion, leading to brittle and unsafe behavior when the robot encounters unexpected contacts. In contrast, SoftMimic enables robots to respond compliantly to external forces while maintaining balance and posture. Our approach leverages an inverse kinematics solver to generate an augmented dataset of feasible compliant motions, which we use to train a reinforcement learning policy. By rewarding the policy for matching compliant responses rather than rigidly tracking the reference motion, SoftMimic learns to absorb disturbances and generalize to varied tasks from a single motion clip. I. INTRODUCTION A major goal in humanoid robotics is to build agents capable of performing a vast range of tasks humans execute in everyday environments. A promising avenue towards this goal is to leverage large-scale human motion capture data, enabling robots to learn human-like behaviors through imitation [1]. All authors are with the Improbable AI Lab, Massachusetts Institute of Technology, USA.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found