FedHybrid: Breaking the Memory Wall of Federated Learning via Hybrid Tensor Management
Tam, Kahou, Tian, Chunlin, Li, Li, Zhao, Haikai, Xu, ChengZhong
–arXiv.org Artificial Intelligence
Federated Learning (FL) emerges as a new learning paradigm that enables multiple devices to collaboratively train a shared model while preserving data privacy. However, one fundamental and prevailing challenge that hinders the deployment of FL on mobile devices is the memory limitation. This paper proposes \textit{FedHybrid}, a novel framework that effectively reduces the memory footprint during the training process while guaranteeing the model accuracy and the overall training progress. Specifically, \textit{FedHybrid} first selects the participating devices for each training round by jointly evaluating their memory budget, computing capability, and data diversity. After that, it judiciously analyzes the computational graph and generates an execution plan for each selected client in order to meet the corresponding memory budget while minimizing the training delay through employing a hybrid of recomputation and compression techniques according to the characteristic of each tensor. During the local training process, \textit{FedHybrid} carries out the execution plan with a well-designed activation compression technique to effectively achieve memory reduction with minimum accuracy loss. We conduct extensive experiments to evaluate \textit{FedHybrid} on both simulation and off-the-shelf mobile devices. The experiment results demonstrate that \textit{FedHybrid} achieves up to a 39.1\% increase in model accuracy and a 15.5$\times$ reduction in wall clock time under various memory budgets compared with the baselines.
arXiv.org Artificial Intelligence
Oct-14-2025
- Country:
- Asia
- China > Zhejiang Province
- Hangzhou (0.05)
- Macao (0.05)
- China > Zhejiang Province
- Europe > United Kingdom (0.04)
- North America > Canada (0.04)
- Asia
- Genre:
- Research Report > New Finding (0.34)
- Industry:
- Information Technology > Security & Privacy (0.68)
- Technology:
- Information Technology
- Artificial Intelligence > Machine Learning
- Neural Networks > Deep Learning (0.46)
- Communications > Mobile (1.00)
- Hardware (1.00)
- Artificial Intelligence > Machine Learning
- Information Technology