Reducing Retraining by Recycling Parameter-Efficient Prompts
Lester, Brian, Yurtsever, Joshua, Shakeri, Siamak, Constant, Noah
–arXiv.org Artificial Intelligence
Parameter-efficient methods are able to use a single frozen pre-trained large language model (LLM) to perform many tasks by learning task-specific soft prompts that modulate model behavior when concatenated to the input text. However, these learned prompts are tightly coupled to a given frozen model -- if the model is updated, corresponding new prompts need to be obtained. In this work, we propose and investigate several approaches to "Prompt Recycling'" where a prompt trained on a source model is transformed to work with the new target model. Our methods do not rely on supervised pairs of prompts, task-specific data, or training updates with the target model, which would be just as costly as re-tuning prompts with the target model from scratch. We show that recycling between models is possible (our best settings are able to successfully recycle $88.9\%$ of prompts, producing a prompt that out-performs baselines), but significant performance headroom remains, requiring improved recycling techniques.
arXiv.org Artificial Intelligence
Aug-10-2022
- Country:
- Asia
- Japan > Honshū
- Chūbu > Toyama Prefecture
- Toyama (0.04)
- Tōhoku > Fukushima Prefecture
- Fukushima (0.04)
- Chūbu > Toyama Prefecture
- Middle East > Jordan (0.04)
- Japan > Honshū
- Europe
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Romania > Sud - Muntenia Development Region
- Giurgiu County > Giurgiu (0.04)
- Belgium > Brussels-Capital Region
- North America > United States
- California > San Diego County
- San Diego (0.04)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- Oregon (0.04)
- Washington > King County
- Seattle (0.14)
- Wisconsin > Dane County
- Madison (0.04)
- California > San Diego County
- Asia
- Genre:
- Research Report (1.00)
- Industry:
- Leisure & Entertainment (0.46)
- Technology: