Stability of Stochastic Gradient Descent on Nonsmooth Convex Losses
–Neural Information Processing Systems
Successful applications of a machine learning algorithm require the algorithm to generalize well to unseen data. Thus understanding and bounding the generalization error of machine learning algorithms is an area of intense theoretical interest and practical importance. The single most popular approach to modern machine learning relies on the use of continuous optimization techniques to optimize the appropriate loss function, most notably the stochastic (sub)gradient descent (SGD) method. Yet the generalization properties of SGD are still not well understood. Consider the setting of stochastic convex optimization (SCO).
Neural Information Processing Systems
May-28-2025, 21:18:30 GMT
- Country:
- North America > United States (0.67)
- Genre:
- Research Report (0.46)
- Industry:
- Information Technology > Security & Privacy (0.46)
- Technology: