abstraction selection
van Seijen
This paper introduces a novel approach for abstraction selection in reinforcement learning problems modelled as factored Markov decision processes (MDPs), for which a state is described via a set of state components. In abstraction selection, an agent must choose an abstraction from a set of candidate abstractions, each build up from a different combination of state components.
Efficient Skill Learning Using Abstraction Selection
Konidaris, George (University of Massachusetts Amherst) | Barto, Andrew (University of Massachusetts Amherst)
We present an algorithm for selecting an appropriate abstraction when learning a new skill. We show empirically that it can consistently select an appropriate abstraction using very little sample data, and that it significantly improves skill learning performance in a reasonably large real-valued reinforcement learning domain.
- North America > United States > Massachusetts > Hampshire County > Amherst (0.05)
- North America > United States > New York > New York County > New York City (0.04)