Emergence of Hierarchy via Reinforcement Learning Using a Multiple Timescale Stochastic RNN
Han, Dongqi, Doya, Kenji, Tani, Jun
Although recurrent neural networks (RNNs) for reinforcement learning (RL) have addressed unique advantages in various aspects, e. g., solving memory-dependent tasks and meta-learning, very few studies have demonstrated how RNNs can solve the problem of hierarchical RL by autonomously developing hierarchical control. In this paper, we propose a novel model-free RL framework called ReMASTER, which combines an off-policy actor-critic algorithm with a multiple timescale stochastic recurrent neural network for solving memory-dependent and hierarchical tasks. We performed experiments using a challenging continuous control task and showed that: (1) Internal representation necessary for achieving hierarchical control autonomously develops through exploratory learning. (2) Stochastic neurons in RNNs enable faster relearning when adapting to a new task which is a recomposition of sub-goals previously learned.
Feb-12-2019
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Education (0.67)
- Health & Medicine > Therapeutic Area
- Neurology (0.46)
- Leisure & Entertainment (0.68)
- Technology: