HSplitLoRA: A Heterogeneous Split Parameter-Efficient Fine-Tuning Framework for Large Language Models
Lin, Zheng, Zhang, Yuxin, Chen, Zhe, Fang, Zihan, Chen, Xianhao, Vepakomma, Praneeth, Ni, Wei, Luo, Jun, Gao, Yue
–arXiv.org Artificial Intelligence
--Recently, large language models (LLMs) have achieved remarkable breakthroughs, revolutionizing the natural language processing domain and beyond. Due to immense parameter sizes, fine-tuning these models with private data for diverse downstream tasks has become mainstream. Though federated learning (FL) offers a promising solution for fine-tuning LLMs without sharing raw data, substantial computing costs hinder its democratization. Moreover, in real-world scenarios, private client devices often possess heterogeneous computing resources, further complicating LLM fine-tuning. T o combat these challenges, we propose HSplitLoRA, a heterogeneous parameter-efficient fine-tuning (PEFT) framework built on split learning (SL) and low-rank adaptation (LoRA) fine-tuning, for efficiently fine-tuning LLMs on heterogeneous client devices. HSplitLoRA first identifies important weights based on their contributions to LLM training. It then dynamically configures the decomposition ranks of LoRA adapters for selected weights and determines the model split point according to varying computing budgets of client devices. Finally, a noise-free adapter aggregation mechanism is devised to support heterogeneous adapter aggregation without introducing noise. Extensive experiments demonstrate that HSplitLoRA outperforms state-of-the-art benchmarks in training accuracy and convergence speed. Index T erms --Distributed learning, split learning, large language model, parameter-efficient fine-tuning. Recently, large language models (LLMs) have achieved tremendous success across a broad spectrum of pivotal sectors due to their exceptional ability in handling high-complexity and large-scale datasets [1]-[5]. Gao are with the Institute of Space Internet, Fudan University, Shanghai 200438, China, and the School of Computer Science, Fudan University, Shanghai 200438, China (email: zlin20@fudan.edu.cn; Z. Lin is also with the Department of Electrical and Electronic Engineering, University of Hong Kong, Pok Fu Lam, Hong Kong, China. X. Chen is with the Department of Electrical and Electronic Engineering, University of Hong Kong, Pok Fu Lam, Hong Kong, China (e-mail: xchen@eee.hku.hk). V epakomma is with Mohamed bin Zayed University of Artificial Intelligence, Abu Dhabi, United Arab Emirates, and the Massachusetts Institute of Technology, Cambridge, MA 02139 USA (e-mail: vepakom@mit.edu).
arXiv.org Artificial Intelligence
May-6-2025
- Country:
- Asia
- China
- Middle East > UAE
- Abu Dhabi Emirate > Abu Dhabi (0.24)
- Singapore (0.04)
- Europe > Romania
- North America > United States
- Massachusetts > Middlesex County > Cambridge (0.54)
- Oceania > Australia
- New South Wales (0.04)
- Asia
- Genre:
- Research Report > Promising Solution (0.34)
- Industry:
- Health & Medicine (0.67)
- Information Technology > Security & Privacy (0.48)
- Technology: