Towards Robust Low-Resource Fine-Tuning with Multi-View Compressed Representations
Liu, Linlin, Li, Xingxuan, Thakkar, Megh, Li, Xin, Joty, Shafiq, Si, Luo, Bing, Lidong
–arXiv.org Artificial Intelligence
Due to the huge amount of parameters, fine-tuning of pretrained language models (PLMs) is prone to overfitting in the low resource scenarios. In this work, we present a novel method that operates on the hidden representations of a PLM to reduce overfitting. During fine-tuning, our method inserts random autoencoders between the hidden layers of a PLM, which transform activations from the previous layers into multi-view compressed representations before feeding them into the upper layers. The autoencoders are plugged out after fine-tuning, so our method does not add extra parameters or increase computation cost during inference. Our method demonstrates promising performance improvement across a wide range of sequence- and token-level low-resource NLP tasks.
arXiv.org Artificial Intelligence
May-26-2023
- Country:
- Europe (1.00)
- North America > United States
- Minnesota (0.28)
- Genre:
- Research Report (0.70)
- Technology: