Two Time-scale Off-Policy TD Learning: Non-asymptotic Analysis over Markovian Samples
Tengyu Xu, Shaofeng Zou, Yingbin Liang
–Neural Information Processing Systems
Gradient-based temporal difference (GTD) algorithms are widely used in off-policy learning scenarios. Among them, the two time-scale TD with gradient correction (TDC) algorithm has been shown to have superior performance. In contrast to previous studies that characterized the non-asymptotic convergence rate of TDC only under identical and independently distributed (i.i.d.) data samples, we provide the first non-asymptotic convergence analysis for two time-scale TDC under a non-i.i.d.
Neural Information Processing Systems
Mar-27-2025, 02:47:36 GMT
- Country:
- North America
- Canada (0.46)
- United States > Ohio (0.40)
- North America
- Genre:
- Research Report > New Finding (0.47)
- Technology: