Stochastic Frank-Wolfe for Constrained Finite-Sum Minimization
Négiar, Geoffrey, Dresdner, Gideon, Tsai, Alicia, Ghaoui, Laurent El, Locatello, Francesco, Freund, Robert M., Pedregosa, Fabian
–arXiv.org Artificial Intelligence
We propose a novel Stochastic Frank-Wolfe (a.k.a. conditional gradient) algorithm for constrained smooth finite-sum minimization with a generalized linear prediction/structure. This class of problems includes empirical risk minimization with sparse, low-rank, or other structured constraints. The proposed method is simple to implement, does not require step-size tuning, and has a constant per-iteration cost that is independent of the dataset size. Furthermore, as a byproduct of the method we obtain a stochastic estimator of the Frank-Wolfe gap that can be used as a stopping criterion. Depending on the setting, the proposed method matches or improves on the best computational guarantees for Stochastic Frank-Wolfe algorithms. Benchmarks on several datasets highlight different regimes in which the proposed method exhibits a faster empirical convergence than related methods. Finally, we provide an implementation of all considered methods in an open-source package.
arXiv.org Artificial Intelligence
Sep-8-2022
- Country:
- Asia > Russia (0.04)
- Europe
- Germany > Baden-Württemberg
- Tübingen Region > Tübingen (0.04)
- Russia (0.04)
- Switzerland > Zürich
- Zürich (0.14)
- Germany > Baden-Württemberg
- North America > United States
- California > Alameda County > Berkeley (0.14)
- Genre:
- Research Report (0.40)
- Workflow (0.46)
- Technology: