Termination Approximation: Continuous State Decomposition for Hierarchical Reinforcement Learning
Harris, Sean (University of New South Wales) | Hengst, Bernhard (University of New South Wales) | Pagnucco, Maurice (University of New South Wales)
This paper presents a divide-and-conquer decomposition for solving continuous state reinforcement learning problems. The contribution lies in a method for stitching together continuous state subtasks in a near-seamless manner along wide continuous boundaries. We introduce the concept of Termination Approximation where the set of subtask termination states are covered by goal sets to generate a set of subtask option policies. The approach employs hierarchical reinforcement learning methods and exploits any underlying repetition in continuous problems to allow reuse of the option policies both within a problem and across related problems. The approach is illustrated using a series of challenging racecar problems.
Mar-1-2015