Convergent Combinations of Reinforcement Learning with Linear Function Approximation
Schoknecht, Ralf, Merke, Artur
–Neural Information Processing Systems
Convergence for iterative reinforcement learning algorithms like TD(O) depends on the sampling strategy for the transitions. However, in practical applications it is convenient to take transition data from arbitrary sources without losing convergence. In this paper we investigate the problem of repeated synchronous updates based on a fixed set of transitions. Our main theorem yields sufficient conditions of convergence for combinations of reinforcement learning algorithms and linear function approximation. This allows to analyse if a certain reinforcement learning algorithm and a certain function approximator are compatible.
Neural Information Processing Systems
Dec-31-2003