Hierarchical Memory-Based Reinforcement Learning
Hernandez-Gardiol, Natalia, Mahadevan, Sridhar
–Neural Information Processing Systems
A key challenge for reinforcement learning is scaling up to large partially observable domains. In this paper, we show how a hierarchy ofbehaviors can be used to create and select among variable length short-term memories appropriate for a task. At higher levels inthe hierarchy, the agent abstracts over lower-level details and looks back over a variable number of high-level decisions in time. We formalize this idea in a framework called Hierarchical Suffix Memory (HSM). HSM uses a memory-based SMDP learning method to rapidly propagate delayed reward across long decision sequences.
Neural Information Processing Systems
Dec-31-2001
- Country:
- North America > United States
- Massachusetts > Middlesex County
- Cambridge (0.14)
- Michigan > Ingham County (0.14)
- Massachusetts > Middlesex County
- North America > United States
- Technology: