Convergent Combinations of Reinforcement Learning with Linear Function Approximation

Schoknecht, Ralf, Merke, Artur

Neural Information Processing Systems 

Convergence for iterative reinforcement learning algorithms like TD(O) depends on the sampling strategy for the transitions. However, in practical applications it is convenient to take transition data from arbitrary sources without losing convergence. In this paper we investigate the problem of repeated synchronous updates based on a fixed set of transitions. Our main theorem yields sufficient conditions of convergence for combinations of reinforcement learning algorithms and linear function approximation. This allows to analyse if a certain reinforcement learning algorithm and a certain function approximator are compatible.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found