Implicit Regularization in Deep Learning May Not Be Explainable by Norms

Neural Information Processing Systems 

Mathematically characterizing the implicit regularization induced by gradient-based optimization is a longstanding pursuit in the theory of deep learning. A widespread hope is that a characterization based on minimization of norms may apply, and a standard test-bed for studying this prospect is matrix factorization (matrix completion via linear neural networks).

Similar Docs  Excel Report  more

TitleSimilaritySource
None found