Continuous-Time Model-Based Reinforcement Learning
Yıldız, Çağatay, Heinonen, Markus, Lähdesmäki, Harri
Model-based reinforcement learning (MBRL) approaches rely on discrete-time state transition models whereas physical systems and the vast majority of control tasks operate in continuous-time. To avoid time-discretization approximation of the underlying process, we propose a continuous-time MBRL framework based on a novel actor-critic method. Our approach also infers the unknown state evolution differentials with Bayesian neural ordinary differential equations (ODE) to account for epistemic uncertainty. We implement and test our method on a new ODE-RL suite that explicitly solves continuous-time control systems. Our experiments illustrate that the model is robust against irregular and noisy data, is sample-efficient, and can solve control problems which pose challenges to discrete-time MBRL methods.
Feb-10-2021
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe
- Finland (0.04)
- Germany > Baden-Württemberg
- Freiburg (0.04)
- North America > United States
- Massachusetts (0.04)
- Asia > Middle East
- Genre:
- Research Report (0.64)
- Technology: