A Single-Step, Sharpness-Aware Minimization is All You Need to Achieve Efficient and Accurate Sparse Training
–Neural Information Processing Systems
However, the training of a sparse DNN encounters great challenges in achieving optimal generalization ability despite the efforts from the state-of-the-art sparse training methodologies. To unravel the mysterious reason behind the difficulty of sparse training, we connect network sparsity with the structure of neural loss functions and identify that the cause of such difficulty lies in a chaotic loss surface.
Neural Information Processing Systems
Feb-12-2026, 20:46:23 GMT
- Country:
- Asia > China
- Tianjin Province > Tianjin (0.04)
- Europe > Italy (0.04)
- North America > United States (0.46)
- Asia > China
- Genre:
- Research Report > Experimental Study (0.93)
- Industry:
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning > Neural Networks (1.00)
- Representation & Reasoning (0.68)
- Vision (0.68)
- Information Technology > Artificial Intelligence