GRID: Scalable Task-Agnostic Prompt-Based Continual Learning for Language Models
Tiwari, Anushka, Pal, Sayantan, Srihari, Rohini K., Ji, Kaiyi
–arXiv.org Artificial Intelligence
Prompt-based continual learning (CL) provides a parameter-efficient approach for adapting large language models (LLMs) across task sequences. However, most existing methods rely on task-aware inference and maintain a growing set of task-specific prompts, which introduces two major challenges: (1) severe performance degradation on earlier tasks under task-agnostic inference, and (2) limited scalability due to prompt memory accumulation as task sequences grow. In this paper, we present GRID, a unified framework designed to address these challenges. GRID incorporates a decoding mechanism that enhances backward transfer by leveraging representative inputs, automatic task identification, and constrained decoding. Furthermore, it employs a gradient-guided prompt selection strategy to compress less informative prompts into a single aggregated representation, ensuring scalable and memory-efficient continual learning. Extensive experiments on long-sequence and negative transfer benchmarks show that GRID improves average accuracy and backward transfer, achieves competitive forward transfer, and substantially reduces prompt memory usage.
arXiv.org Artificial Intelligence
Oct-2-2025
- Country:
- Asia > Thailand
- Europe > Ireland
- Leinster > County Dublin > Dublin (0.04)
- North America
- Dominican Republic (0.04)
- United States > Florida
- Miami-Dade County > Miami (0.04)
- South America > Colombia
- Meta Department > Villavicencio (0.04)
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Education (0.46)
- Technology: