3-in-1: 2D Rotary Adaptation for Efficient Finetuning, Efficient Batching and Composability
–Neural Information Processing Systems
Parameter-efficient finetuning (PEFT) methods effectively adapt large language models (LLMs) to diverse downstream tasks, reducing storage and GPU memory demands. Despite these advantages, several applications pose new challenges to PEFT beyond mere parameter efficiency. One notable challenge involves the efficient deployment of LLMs equipped with multiple task- or user-specific adapters, particularly when different adapters are needed for distinct requests within the same batch. Another challenge is the interpretability of LLMs, which is crucial for understanding how LLMs function. Previous studies introduced various approaches to address different challenges.
Neural Information Processing Systems
May-26-2025, 22:11:34 GMT
- Technology: