Parameter-Efficient Multilingual Summarisation: An Empirical Study
Whitehouse, Chenxi, Huot, Fantine, Bastings, Jasmijn, Dehghani, Mostafa, Lin, Chu-Cheng, Lapata, Mirella
–arXiv.org Artificial Intelligence
With the increasing prevalence of Large Language Models, traditional full fine-tuning approaches face growing challenges, especially in memory-intensive tasks. This paper investigates the potential of Parameter-Efficient Fine-Tuning, focusing on Low-Rank Adaptation (LoRA), for complex and under-explored multilingual summarisation tasks. We conduct an extensive study across different data availability scenarios, including full-data, low-data, and cross-lingual transfer, leveraging models of different sizes. Our findings reveal that LoRA lags behind full fine-tuning when trained with full data, however, it excels in low-data scenarios and cross-lingual transfer. Interestingly, as models scale up, the performance gap between LoRA and full fine-tuning diminishes. Additionally, we investigate effective strategies for few-shot cross-lingual transfer, finding that continued LoRA tuning achieves the best performance compared to both full fine-tuning and dynamic composition of language-specific LoRA modules.
arXiv.org Artificial Intelligence
Nov-14-2023
- Country:
- Asia
- China (0.14)
- Middle East > UAE (0.14)
- Europe
- North America
- Canada (0.14)
- United States (0.14)
- Asia
- Genre:
- Research Report > New Finding (0.34)
- Industry:
- Education > Educational Setting > Continuing Education (0.30)
- Technology: