Forward and Backward State Abstractions for Off-policy Evaluation

Hao, Meiling, Su, Pingfan, Hu, Liyuan, Szabo, Zoltan, Zhao, Qingyuan, Shi, Chengchun

arXiv.org Machine Learning 

Off-policy evaluation (OPE) is crucial for evaluating a target policy's impact offline before its deployment. However, achieving accurate OPE in large state spaces remains challenging. This paper studies state abstractions - originally designed for policy learning - in the context of OPE. Our contributions are three-fold: (i) We define a set of irrelevance conditions central to learning state abstractions for OPE.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found