Efficient Neural Network Training via Forward and Backward Propagation Sparsification
–Neural Information Processing Systems
Sparse training is a natural idea to accelerate the training speed of deep neural networks and save the memory usage, especially since large modern neural networks are significantly over-parameterized. However, most of the existing methods cannot achieve this goal in practice because the chain rule based gradient (w.r.t.
Neural Information Processing Systems
Feb-9-2026, 13:45:56 GMT
- Technology: