VB-LoRA: Extreme Parameter Efficient Fine-Tuning with Vector Banks
Li, Yang, Han, Shaobo, Ji, Shihao
–arXiv.org Artificial Intelligence
As the adoption of large language models increases and the need for per-user or per-task model customization grows, the parameter-efficient fine-tuning (PEFT) methods, such as low-rank adaptation (LoRA) and its variants, incur substantial storage and transmission costs. To further reduce stored parameters, we introduce a "divide-and-share" paradigm that breaks the barriers of low-rank decomposition across matrix dimensions, modules and layers by sharing parameters globally via a vector bank. As an instantiation of the paradigm to LoRA, our proposed VB-LoRA composites all the low-rank matrices of LoRA from a shared vector bank with a differentiable top-$k$ admixture module. VB-LoRA achieves extreme parameter efficiency while maintaining comparable or better performance compared to state-of-the-art PEFT methods. Extensive experiments demonstrate the effectiveness of VB-LoRA on natural language understanding, natural language generation, and instruction tuning tasks. When fine-tuning the Llama2-13B model, VB-LoRA only uses 0.4% of LoRA's stored parameters, yet achieves superior results. Our source code is available at https://github.com/leo-yangli/VB-LoRA.
arXiv.org Artificial Intelligence
May-27-2024
- Country:
- Africa > Senegal
- Kolda Region > Kolda (0.04)
- Europe
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- Germany > Saarland
- Saarbrücken (0.04)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Romania > Sud - Muntenia Development Region
- Giurgiu County > Giurgiu (0.04)
- Belgium > Brussels-Capital Region
- North America
- Canada > British Columbia
- United States
- Arizona (0.04)
- Georgia > Fulton County
- Atlanta (0.04)
- Hawaii > Honolulu County
- Honolulu (0.04)
- New Jersey > Mercer County
- Princeton (0.04)
- Washington > King County
- Seattle (0.04)
- Oceania > Australia
- South America > Colombia
- Meta Department > Villavicencio (0.04)
- Africa > Senegal
- Genre:
- Overview (0.68)
- Research Report (0.82)
- Technology: