Pre-Training and Personalized Fine-Tuning via Over-the-Air Federated Meta-Learning: Convergence-Generalization Trade-Offs
Wen, Haifeng, Xing, Hong, Simeone, Osvaldo
–arXiv.org Artificial Intelligence
For modern artificial intelligence (AI) applications such as large language models (LLMs), the training paradigm has recently shifted to pre-training followed by fine-tuning. Furthermore, owing to dwindling open repositories of data and thanks to efforts to democratize access to AI models, pre-training is expected to increasingly migrate from the current centralized deployments to federated learning (FL) implementations. Meta-learning provides a general framework in which pre-training and fine-tuning can be formalized. Meta-learning-based personalized FL (meta-pFL) moves beyond basic personalization by targeting generalization to new agents and tasks. This paper studies the generalization performance of meta-pFL for a wireless setting in which the agents participating in the pre-training phase, i.e., meta-learning, are connected via a shared wireless channel to the server. Adopting over-the-air computing, we study the trade-off between generalization to new agents and tasks, on the one hand, and convergence, on the other hand. The trade-off arises from the fact that channel impairments may enhance generalization, while degrading convergence. Extensive numerical results validate the theory.
arXiv.org Artificial Intelligence
Jun-17-2024
- Country:
- Africa > Rwanda
- Asia
- Malaysia > Kuala Lumpur
- Kuala Lumpur (0.04)
- Singapore (0.04)
- Malaysia > Kuala Lumpur
- Europe
- Greece (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- North America
- Canada
- British Columbia > Metro Vancouver Regional District
- Vancouver (0.04)
- Quebec > Montreal (0.04)
- British Columbia > Metro Vancouver Regional District
- United States
- Massachusetts (0.04)
- Rhode Island (0.04)
- Canada
- Oceania > Australia
- New South Wales > Sydney (0.04)
- Genre:
- Research Report (1.00)
- Technology: