When to Trust Your Model: Model-Based Policy Optimization
Michael Janner, Justin Fu, Marvin Zhang, Sergey Levine
–Neural Information Processing Systems
Designing effective model-based reinforcement learning algorithms is difficult because the ease of data generation must be weighed against the bias of model-generated data. In this paper, we study the role of model usage in policy optimization both theoretically and empirically.
Neural Information Processing Systems
Nov-16-2025, 17:52:27 GMT
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe > United Kingdom
- England > Cambridgeshire > Cambridge (0.04)
- North America
- Canada (0.04)
- United States > California
- Alameda County > Berkeley (0.04)
- Asia > Middle East
- Genre:
- Research Report (0.46)
- Technology: