Actor-Expert: A Framework for using Action-Value Methods in Continuous Action Spaces
Lim, Sungsu, Joseph, Ajin, Le, Lei, Pan, Yangchen, White, Martha
–arXiv.org Artificial Intelligence
Value-based approaches can be difficult to use in continuous action spaces, because an optimization has to be solved to find the greedy action for the action-values. A common strategy has been to restrict the functional form of the action-values to be convex or quadratic in the actions, to simplify this optimization. Such restrictions, however, can prevent learning accurate action-values. In this work, we propose the Actor-Expert framework for value-based methods, that decouples action-selection (Actor) from the action-value representation (Expert). The Expert uses Q-learning to update the action-values towards the optimal action-values, whereas the Actor (learns to) output the greedy action for the current action-values. We develop a Conditional Cross Entropy Method for the Actor, to learn the greedy action for a generically parameterized Expert, and provide a two-timescale analysis to validate asymptotic behavior. We demonstrate in a toy domain with bimodal action-values that previous restrictive action-value methods fail whereas the decoupled Actor-Expert with a more general action-value parameterization succeeds. Finally, we demonstrate that Actor-Expert performs as well as or better than these other methods on several benchmark continuous-action domains.
arXiv.org Artificial Intelligence
Oct-22-2018
- Country:
- North America
- Canada > Alberta (0.14)
- United States (0.28)
- North America
- Genre:
- Research Report (0.63)
- Technology: