The Curious Price of Distributional Robustness in Reinforcement Learning with a Generative Model

Neural Information Processing Systems 

This paper investigates model robustness in reinforcement learning (RL) via the framework of distributionally robust Markov decision processes (RMDPs). Despite recent efforts, the sample complexity of RMDPs is much less understood regardless of the uncertainty set in use; in particular, there exist large gaps between existing upper and lower bounds, and it is unclear if distributional robustness bears any statistical implications when benchmarked against standard RL.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found