Manifold Embeddings for Model-Based Reinforcement Learning under Partial Observability
–Neural Information Processing Systems
Interesting real-world datasets often exhibit nonlinear, noisy, continuous-valued states that are unexplorable, are poorly described by first principles, and are only partially observable. If partial observability can be overcome, these constraints suggest the use of model-based reinforcement learning. We experiment with manifold embeddings as the reconstructed observable state-space of an off-line, model-based reinforcement learning approach to control. We demonstrate the embedding of a system changes as a result of learning and that the best performing embeddings well-represent the dynamics of both the uncontrolled and adaptively controlled system. We apply this approach in simulation to learn a neurostimulation policy that is more efficient in treating epilepsy than conventional policies. We then demonstrate the learned policy completely suppressing seizures in real-world neurostimulation experiments on actual animal brain slices.
Neural Information Processing Systems
Dec-31-2009
- Country:
- North America > Canada > Quebec > Montreal (0.14)
- Industry:
- Health & Medicine > Therapeutic Area > Neurology > Epilepsy (0.49)
- Technology: