Goto

Collaborating Authors

 Ahmet Alacaoglu



An Inexact Augmented Lagrangian Framework for Nonconvex Optimization with Nonlinear Constraints

Neural Information Processing Systems

We propose a practical inexact augmented Lagrangian method (iALM) for nonconvex problems with nonlinear constraints. We characterize the total computational complexity of our method subject to a verifiable geometric condition, which is closely related to the Polyak-Lojasiewicz and Mangasarian-Fromowitz conditions.


Smooth Primal-Dual Coordinate Descent Algorithms for Nonsmooth Convex Optimization

Neural Information Processing Systems

We propose a new randomized coordinate descent method for a convex optimization template with broad applications. Our analysis relies on a novel combination of four ideas applied to the primal-dual gap function: smoothing, acceleration, homotopy, and coordinate descent with non-uniform sampling. As a result, our method features the first convergence rate guarantees among the coordinate descent methods, that are the best-known under a variety of common structure assumptions on the template. We provide numerical evidence to support the theoretical results with a comparison to state-of-the-art algorithms.