Model-Free Active Exploration in Reinforcement Learning

Neural Information Processing Systems 

We study the problem of exploration in Reinforcement Learning and present a novel model-free solution. We adopt an information-theoretical viewpoint and start from the instance-specific lower bound of the number of samples that have to be collected to identify a nearly-optimal policy.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found