A Concise Lyapunov Analysis of Nesterov's Accelerated Gradient Method
–arXiv.org Artificial Intelligence
Among them, Nesterov's accelerated gradient method [7,8] has gained significant attention due to its provable acceleration on general convex functions beyond quadratics. A special focus has been on using dynamical system tools [12,10,3,14] and control-theoretical methods [5,9] for the analysis and design of such algorithms. In the standard textbook [8] by Nesterov, the convergence analysis of accelerated gradient methods is conducted using a technique known as estimating sequences. These are essentially auxiliary comparison functions used to prove the convergence rates of optimization algorithms. As pointed out in [14], estimating sequences are usually constructed inductively and can be difficult to understand and apply. This motivated the Lyapunov analysis in [14], which aims to unify the analysis of a broad class of accelerated algorithms. Despite this comprehensive work, to the best knowledge of the author, a simple and direct Lyapunov analysis of the original scheme of Nesterov's accelerated gradient method is still lacking.
arXiv.org Artificial Intelligence
Feb-25-2025
- Country:
- Asia
- Middle East > Jordan (0.04)
- Russia (0.04)
- Europe > Russia (0.04)
- North America > Canada
- Ontario > Waterloo Region > Waterloo (0.04)
- Asia
- Genre:
- Research Report (0.40)
- Technology: