Decoupling Representation Learning from Reinforcement Learning
Stooke, Adam, Lee, Kimin, Abbeel, Pieter, Laskin, Michael
–arXiv.org Artificial Intelligence
In an effort to overcome limitations of reward-driven feature learning in deep reinforcement learning (RL) from images, we propose decoupling representation learning from policy learning. To this end, we introduce a new unsupervised learning (UL) task, called Augmented Temporal Contrast (ATC), which trains a convolutional encoder to associate pairs of observations separated by a short time difference, under image augmentations and using a contrastive loss. In online RL experiments, we show that training the encoder exclusively using ATC matches or outperforms end-to-end RL in most environments. Additionally, we benchmark several leading UL algorithms by pre-training encoders on expert demonstrations and using them, with weights frozen, in RL agents; we find that agents using ATC-trained encoders outperform all others. We also train multi-task encoders on data from multiple environments and show generalization to different downstream RL tasks. Finally, we ablate components of ATC, and introduce a new data augmentation to enable replay of (compressed) latent images from pre-trained encoders when RL requires augmentation. Ever since the first fully-learned approach succeeded at playing Atari games from screen images (Mnih et al., 2015), standard practice in deep reinforcement learning (RL) has been to learn visual features and a control policy jointly, end-to-end.
arXiv.org Artificial Intelligence
Sep-14-2020
- Country:
- North America > United States > California > Alameda County > Berkeley (0.14)
- Genre:
- Research Report > New Finding (0.67)
- Industry:
- Leisure & Entertainment > Games > Computer Games (0.70)
- Technology: