Advancing Model Pruning via Bi-level Optimization Parikshit Ram 2 Pu Zhao 3
–Neural Information Processing Systems
As illustrated by the Lottery Ticket Hypothesis (LTH), pruning also has the potential of improving their generalization ability. At the core of LTH, iterative magnitude pruning (IMP) is the predominant pruning method to successfully find'winning tickets'. Yet, the computation cost of IMP grows prohibitively as the targeted pruning ratio increases. To reduce the computation overhead, various efficient'one-shot' pruning methods have been developed but these schemes are usually unable to find winning tickets as good as IMP. This raises the question of how to close the gap between pruning accuracy and pruning efficiency? To tackle it, we pursue the algorithmic advancement of model pruning. Specifically, we formulate the pruning problem from a fresh and novel viewpoint, bi-level optimization (BLO). We show that the BLO interpretation provides a technically-grounded optimization base for an efficient implementation of the pruning-retraining learning paradigm used in IMP.
Neural Information Processing Systems
Mar-26-2025, 04:20:24 GMT
- Country:
- North America > United States (0.67)
- Genre:
- Contests & Prizes (0.79)
- Research Report > New Finding (0.46)
- Industry:
- Leisure & Entertainment (0.93)
- Technology: