Make Sharpness-Aware Minimization Stronger: A Sparsified Perturbation Approach

Neural Information Processing Systems 

Deep neural networks often suffer from poor generalization caused by complex and non-convex loss landscapes. One of the popular solutions is Sharpness-A ware Minimization (SAM), which smooths the loss landscape via minimizing the maximized change of training loss when adding a perturbation to the weight.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found