Accelerating Primal-dual Methods for Regularized Markov Decision Processes
Li, Haoya, Yu, Hsiang-fu, Ying, Lexing, Dhillon, Inderjit
–arXiv.org Artificial Intelligence
Entropy regularized Markov decision processes have been widely used in reinforcement learning. This paper is concerned with the primal-dual formulation of the entropy regularized problems. Standard first-order methods suffer from slow convergence due to the lack of strict convexity and concavity. To address this issue, we first introduce a new quadratically convexified primal-dual formulation. The natural gradient ascent descent of the new formulation enjoys a global convergence guarantee and exponential convergence rate. We also propose a new interpolating metric that further accelerates the convergence significantly. Numerical results are provided to demonstrate the performance of the proposed methods under multiple settings.
arXiv.org Artificial Intelligence
Jun-12-2023
- Country:
- North America > United States
- California > Santa Clara County (0.14)
- Texas (0.14)
- North America > United States
- Genre:
- Research Report (0.65)