Stochastic Gradient Descent, Weighted Sampling, and the Randomized Kaczmarz algorithm
–Neural Information Processing Systems
We improve a recent gurantee of Bach and Moulines on the linear convergence of SGD for smooth and strongly convex objectives, reducing a quadratic dependence on the strong convexity to a linear dependence. Furthermore, we show how reweighting the sampling distribution (i.e.
Neural Information Processing Systems
Sep-30-2025, 10:14:03 GMT
- Technology: