QMDP-Net: Deep Learning for Planning under Partial Observability
Peter Karkus, David Hsu, Wee Sun Lee
–Neural Information Processing Systems
This paper introduces the QMDP-net, a neural network architecture for planning under partial observability. The QMDP-net combines the strengths of model-free learning and model-based planning. It is a recurrent policy network, but it represents a policy for a parameterized set of tasks by connecting a model with a planning algorithm that solves the model, thus embedding the solution structure of planning in a network learning architecture. The QMDP-net is fully differentiable and allows for end-to-end training. We train a QMDPnet on different tasks so that it can generalize to new ones in the parameterized task set and "transfer" to other similar tasks beyond the set. In preliminary experiments, QMDP-net showed strong performance on several robotic tasks in simulation. Interestingly, while QMDP-net encodes the QMDP algorithm, it sometimes outperforms the QMDP algorithm in the experiments, as a result of end-to-end learning.
Neural Information Processing Systems
Oct-4-2024, 09:35:56 GMT
- Country:
- Asia > Singapore (0.04)
- Europe > Germany
- Baden-Württemberg > Freiburg (0.04)
- North America > United States
- California > Los Angeles County > Long Beach (0.04)
- Genre:
- Research Report (0.46)
- Industry:
- Education (0.68)
- Technology: