Forward and Backward State Abstractions for Off-policy Evaluation
Hao, Meiling, Su, Pingfan, Hu, Liyuan, Szabo, Zoltan, Zhao, Qingyuan, Shi, Chengchun
Off-policy evaluation (OPE) is crucial for evaluating a target policy's impact offline before its deployment. However, achieving accurate OPE in large state spaces remains challenging. This paper studies state abstractions - originally designed for policy learning - in the context of OPE. Our contributions are three-fold: (i) We define a set of irrelevance conditions central to learning state abstractions for OPE.
Jun-27-2024
- Country:
- Asia (0.28)
- Europe > United Kingdom
- England > Cambridgeshire > Cambridge (0.14)
- North America > United States
- Massachusetts (0.14)
- Genre:
- Research Report (1.00)
- Industry:
- Health & Medicine (0.46)
- Technology: