LST: Ladder Side-Tuning for Parameter and Memory Efficient Transfer Learning
–Neural Information Processing Systems
Fine-tuning large pre-trained models on downstream tasks has been adopted in a variety of domains recently. However, it is costly to update the entire parameter set of large pre-trained models. Although recently proposed parameter-efficient transfer learning (PETL) techniques allow updating a small subset of parameters (e.g. This is because the gradient computation for the trainable parameters still requires back-propagation through the large pre-trained backbone model. To address this, we propose Ladder Side-Tuning (LST), a new PETL technique that can reduce training memory requirements by more substantial amounts.
Neural Information Processing Systems
Oct-11-2024, 02:36:27 GMT
- Technology: