A Win-win Deal: Towards Sparse and Robust Pre-trained Language Models
–Neural Information Processing Systems
In response to the efficiency problem, recent studies show that dense PLMs can be replaced with sparse subnetworks without hurting the performance. Such subnetworks can be found in three scenarios: 1) the fine-tuned PLMs, 2) the raw PLMs and then fine-tuned in isolation, and even inside 3) PLMs without any parameter fine-tuning. However, these results are only obtained in the in-distribution (ID) setting.
Neural Information Processing Systems
Aug-16-2025, 04:42:01 GMT