Understanding and Addressing the Pitfalls of Bisimulation-based Representations in Offline Reinforcement Learning
–Neural Information Processing Systems
While bisimulation-based approaches hold promise for learning robust state representations for Reinforcement Learning (RL) tasks, their efficacy in offline RL tasks has not been up to par. In some instances, their performance has even significantly underperformed alternative methods. We aim to understand why bisimulation methods succeed in online settings, but falter in offline tasks. Our analysis reveals that missing transitions in the dataset are particularly harmful to the bisimulation principle, leading to ineffective estimation. We also shed light on the critical role of reward scaling in bounding the scale of bisimulation measurements and of the value error they induce.
Neural Information Processing Systems
Jan-18-2025, 13:54:40 GMT
- Technology: