Model-based Reinforcement Learning for Semi-Markov Decision Processes with Neural ODEs
Du, Jianzhun, Futoma, Joseph, Doshi-Velez, Finale
We present two elegant solutions for modeling continuous-time dynamics, in a novel model-based reinforcement learning (RL) framework for semi-Markov decision processes (SMDPs), using neural ordinary differential equations (ODEs). Our models accurately characterize continuous-time dynamics and enable us to develop high-performing policies using a small amount of data. We also develop a model-based approach for optimizing time schedules to reduce interaction rates with the environment while maintaining the near-optimal performance, which is not possible for model-free methods. We experimentally demonstrate the efficacy of our methods across various continuous-time domains.
Oct-25-2020
- Country:
- North America
- Canada > British Columbia
- United States
- California > Alameda County
- Berkeley (0.14)
- Massachusetts > Middlesex County
- Cambridge (0.04)
- New York (0.04)
- North Carolina (0.04)
- California > Alameda County
- North America
- Genre:
- Research Report (1.00)
- Industry:
- Health & Medicine > Therapeutic Area
- Immunology > HIV (0.48)
- Infections and Infectious Diseases (1.00)
- Health & Medicine > Therapeutic Area
- Technology: