AdaRL: What, Where, and How to Adapt in Transfer Reinforcement Learning
Huang, Biwei, Feng, Fan, Lu, Chaochao, Magliacane, Sara, Zhang, Kun
–arXiv.org Artificial Intelligence
Most approaches in reinforcement learning (RL) are data-hungry and specific to fixed environments. In this paper, we propose a principled framework for adaptive RL, called AdaRL, that adapts reliably to changes across domains. Specifically, we construct a generative environment model for the structural relationships among variables in the system and embed the changes in a compact way, which provides a clear and interpretable picture for locating what and where the changes are and how to adapt. Based on the environment model, we characterize a minimal set of representations, including both domain-specific factors and domain-shared state representations, that suffice for reliable and low-cost transfer. Moreover, we show that by explicitly leveraging a compact representation to encode changes, we can adapt the policy with only a few samples without further policy optimization in the target domain. We illustrate the efficacy of AdaRL through a series of experiments that allow for changes in different components of Cartpole and Atari games.
arXiv.org Artificial Intelligence
Jul-7-2021
- Country:
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.14)
- Genre:
- Research Report (0.50)
- Industry:
- Leisure & Entertainment > Games > Computer Games (0.54)
- Technology: