Goto

Collaborating Authors

 mathematicalprogramming




Smoothed analysis of the low-rank approach for smooth semidefinite programs

Thomas Pumir, Samy Jelassi, Nicolas Boumal

Neural Information Processing Systems

Inprior work, ithas been shown that, when the constraints on the factorized variable regularly define a smooth manifold, providedk is large enough, for almost all cost matrices, all second-order stationary points (SOSPs) are optimal. Importantly, in practice, one can only compute points which approximately satisfy necessary optimality conditions, leading tothequestion: aresuch points also approximately optimal?


Natasha 2: Faster Non-Convex Optimization Than SGD

Zeyuan Allen-Zhu

Neural Information Processing Systems

In diverse world of deep learning research has given rise to numerous architectures for neural networks(convolutionalones,longshorttermmemoryones,etc). However,tothisdate,theunderlying training algorithms for neural networks are still stochastic gradient descent (SGD) and its heuristic variants.