Watch Less, Feel More: Sim-to-Real RL for Generalizable Articulated Object Manipulation via Motion Adaptation and Impedance Control

Do, Tan-Dzung, Gireesh, Nandiraju, Wang, Jilong, Wang, He

arXiv.org Artificial Intelligence 

Watch Less, Feel More: Sim-to-Real RL for Generalizable Articulated Object Manipulation via Motion Adaptation and Impedance Control Tan-Dzung Do 1,2, Nandiraju Gireesh 1,2, Jilong Wang 2, and He Wang 1,2, Figure 1: We train an RL policy to open doors and drawers in simulation that adapts its action according to the motion of objects by leveraging history observations (left). We directly transfer this policy to reach 80% joint limit in the real world with closed-loop variable impedance control and achieve 84% success rate, using only one first-frame RGBD image (right). Abstract -- Articulated object manipulation poses a unique challenge compared to rigid object manipulation as the object itself represents a dynamic environment. In this work, we present a novel RL-based pipeline equipped with variable impedance control and motion adaptation leveraging observation history for generalizable articulated object manipulation, focusing on smooth and dexterous motion during zero-shot sim-to-real transfer (Figure 1). T o mitigate the sim-to-real gap, our pipeline diminishes reliance on vision by not leveraging the vision data feature (RGBD/pointcloud) directly as policy input but rather extracting useful low-dimensional data first via off-the-shelf modules. Additionally, we experience less sim-to-real gap by inferring object motion and its intrinsic properties via observation history as well as utilizing impedance control both in the simulation and in the real world. Furthermore, we develop a well-designed training setting with great randomization and a specialized reward system (task-aware and motion-aware) that enables multi-staged, end-to-end manipulation without heuristic motion planning.