Zero-shot Sim2Real Adaptation Across Environments
Semage, Buddhika Laknath, Karimpanal, Thommen George, Rana, Santu, Venkatesh, Svetha
–arXiv.org Artificial Intelligence
Simulation based learning often provides a cost-efficient recourse to reinforcement learning applications in robotics. However, simulators are generally incapable of accurately replicating real-world dynamics, and thus bridging the sim2real gap is an important problem in simulation based learning. Current solutions to bridge the sim2real gap involve hybrid simulators that are augmented with neural residual models. Unfortunately, they require a separate residual model for each individual environment configuration (i.e., a fixed setting of environment variables such as mass, friction etc.), and thus are not transferable to new environments quickly. To address this issue, we propose a Reverse Action Transformation (RAT) policy which learns to imitate simulated policies in the real-world. Once learnt from a single environment, RAT can then be deployed on top of a Universal Policy Network to achieve zero-shot adaptation to new environments. We empirically evaluate our approach in a set of continuous control tasks and observe its advantage as a few-shot and zero-shot learner over competing baselines.
arXiv.org Artificial Intelligence
Feb-8-2023
- Country:
- Europe
- France (0.04)
- Germany > Baden-Württemberg
- Freiburg (0.04)
- North America
- Canada > Quebec
- Montreal (0.04)
- United States
- Massachusetts > Middlesex County
- Cambridge (0.04)
- New York > New York County
- New York City (0.04)
- Pennsylvania > Allegheny County
- Pittsburgh (0.04)
- Massachusetts > Middlesex County
- Canada > Quebec
- Oceania > Australia (0.04)
- Europe
- Genre:
- Research Report (0.64)
- Workflow (0.68)
- Technology: