Goto

Collaborating Authors

 algorithm and analysis


Stochastic Distributed Optimization under Average Second-order Similarity: Algorithms and Analysis

Neural Information Processing Systems

We study finite-sum distributed optimization problems involving a master node and $n-1$ local nodes under the popular $\delta$-similarity and $\mu$-strong convexity conditions. We propose two new algorithms, SVRS and AccSVRS, motivated by previous works. The non-accelerated SVRS method combines the techniques of gradient sliding and variance reduction and achieves a better communication complexity of $\tilde{\mathcal{O}}(n {+} \sqrt{n}\delta/\mu)$ compared to existing non-accelerated algorithms.


Stochastic Distributed Optimization under Average Second-order Similarity: Algorithms and Analysis

Neural Information Processing Systems

We study finite-sum distributed optimization problems involving a master node and n-1 local nodes under the popular \delta -similarity and \mu -strong convexity conditions. We propose two new algorithms, SVRS and AccSVRS, motivated by previous works. The non-accelerated SVRS method combines the techniques of gradient sliding and variance reduction and achieves a better communication complexity of \tilde{\mathcal{O}}(n { } \sqrt{n}\delta/\mu) compared to existing non-accelerated algorithms. In contrast to existing results, our complexity bounds are entirely smoothness-free and exhibit superiority in ill-conditioned cases. Furthermore, we establish a nearly matched lower bound to verify the tightness of our AccSVRS method.


An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback

Shamir, Ohad

arXiv.org Machine Learning

We consider the closely related problems of bandit convex optimization with two-point feedback, and zero-order stochastic convex optimization with two function evaluations per round. We provide a simple algorithm and analysis which is optimal for convex Lipschitz functions. This improves on \cite{dujww13}, which only provides an optimal result for smooth functions; Moreover, the algorithm and analysis are simpler, and readily extend to non-Euclidean problems. The algorithm is based on a small but surprisingly powerful modification of the gradient estimator.