Sharpness, Restart and Acceleration
Vincent Roulet, Alexandre d'Aspremont
–Neural Information Processing Systems
The Łojasiewicz inequality shows that sharpness bounds on the minimum of convex optimization problems hold almost generically. Sharpness directly controls the performance of restart schemes, as observed by Nemirovskii and Nesterov [1985]. The constants quantifying error bounds are of course unobservable, but we show that optimal restart strategies are robust, and searching for the best scheme only increases the complexity by a logarithmic factor compared to the optimal bound. Overall then, restart schemes generically accelerate accelerated methods.
Neural Information Processing Systems
Oct-7-2024, 16:29:06 GMT