Solving POMDPs by Searching in Policy Space
–arXiv.org Artificial Intelligence
Most algorithms for solving POMDPs iteratively improve a value function that implicitly represents a policy and are said to search in value function space. This paper presents an approach to solving POMDPs that represents a policy explicitly as a finite-state controller and iteratively improves the controller by search in policy space. Two related algorithms illustrate this approach. The first is a policy iteration algorithm that can outperform value iteration in solving infinitehorizon POMDPs. It provides the foundation for a new heuristic search algorithm that promises further speedup by focusing computational effort on regions of the problem space that are reachable, or likely to be reached, from a start state.
arXiv.org Artificial Intelligence
Jan-30-2013
- Country:
- North America > United States
- Massachusetts > Hampshire County
- Amherst (0.14)
- Texas > Travis County
- Austin (0.04)
- Massachusetts > Hampshire County
- North America > United States
- Genre:
- Research Report (0.82)
- Industry: