A Win-win Deal: Towards Sparse and Robust Pre-trained Language Models

Neural Information Processing Systems 

In response to the efficiency problem, recent studies show that dense PLMs can be replaced with sparse subnetworks without hurting the performance. Such subnetworks can be found in three scenarios: 1) the fine-tuned PLMs, 2) the raw PLMs and then fine-tuned in isolation, and even inside 3) PLMs without any parameter fine-tuning. However, these results are only obtained in the in-distribution (ID) setting.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found