RaSA: Rank-Sharing Low-Rank Adaptation
He, Zhiwei, Tu, Zhaopeng, Wang, Xing, Chen, Xingyu, Wang, Zhijie, Xu, Jiahao, Liang, Tian, Jiao, Wenxiang, Zhang, Zhuosheng, Wang, Rui
–arXiv.org Artificial Intelligence
Low-rank adaptation (LoRA) has been prominently employed for parameterefficient fine-tuning of large language models (LLMs). However, the limited expressive capacity of LoRA, stemming from the low-rank constraint, has been recognized as a bottleneck, particularly in rigorous tasks like code generation and mathematical reasoning. To address this limitation, we introduce Rank-Sharing Low-Rank Adaptation (RaSA), an innovative extension that enhances the expressive capacity of LoRA by leveraging partial rank sharing across layers. By forming a shared rank pool and applying layer-specific weighting, RaSA effectively increases the number of ranks without augmenting parameter overhead. Our theoretically grounded and empirically validated approach demonstrates that RaSA not only maintains the core advantages of LoRA but also significantly boosts performance in challenging code and math tasks. Code, data and scripts are available at: https://github.com/zwhe99/RaSA. Low-rank adaptation (LoRA, Hu et al. (2022)) has become a de facto parameter-efficient fine-tuning (PEFT) method for adapting large language models (LLMs) to specific downstream tasks. Its core idea is to constrain the parameter updates to be low-rank, which significantly reduces the number of trainable parameters and allows them to be merged back into the original model, thereby avoiding additional inference latency.
arXiv.org Artificial Intelligence
Mar-16-2025
- Country:
- Europe > Middle East
- Malta (0.14)
- North America > Mexico
- Mexico City (0.14)
- Europe > Middle East
- Genre:
- Research Report > Promising Solution (0.34)
- Technology: