Implicit Regularization or Implicit Conditioning Exact Risk Trajectories of in High Dimensions

Neural Information Processing Systems 

Stochastic gradient descent (SGD) is a pillar of modern machine learning, serving as the go-to optimization algorithm for a diverse array of problems. While the empirical success of SGD is often attributed to its computational efficiency and favorable generalization behavior, neither effect is well understood and disentangling them remains an open problem.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found