Off-Policy Evaluation via Off-Policy Classification Alex Irpan
–Neural Information Processing Systems
In this work, we consider the problem of model selection for deep reinforcement learning (RL) in real-world environments. Typically, the performance of deep RL algorithms is evaluated via on-policy interactions with the target environment. However, comparing models in a real-world environment for the purposes of early stopping or hyperparameter tuning is costly and often practically infeasible. This leads us to examine off-policy policy evaluation (OPE) in such settings. We focus on OPE for value-based methods, which are of particular interest in deep RL, with applications like robotics, where off-policy algorithms based on Q-function estimation can often attain better sample complexity than direct policy optimization. Existing OPE metrics either rely on a model of the environment, or the use of importance sampling (IS) to correct for the data being off-policy.
Neural Information Processing Systems
Jun-1-2025, 09:58:30 GMT
- Country:
- North America > United States > California (0.14)
- Genre:
- Research Report > New Finding (0.68)
- Technology: