Thinking Forward: Memory-Efficient Federated Finetuning of Language Models
–Neural Information Processing Systems
Finetuning large language models (LLMs) in federated learning (FL) settings has become increasingly important as it allows resource-constrained devices to finetune a model using private data. However, finetuning LLMs using backpropagation requires excessive memory (especially from intermediate activations) for resource-constrained devices. While Forward-mode Auto-Differentiation (AD) can significantly reduce memory footprint from activations, we observe that directly applying it to LLM finetuning results in slow convergence and poor accuracy.
Neural Information Processing Systems
Mar-23-2025, 01:30:53 GMT
- Country:
- North America > United States > Massachusetts > Hampshire County > Amherst (0.14)
- Genre:
- Research Report > Experimental Study (0.92)
- Industry:
- Information Technology > Security & Privacy (0.66)
- Technology: