Continual Learning with Global Alignment
–Neural Information Processing Systems
Continual learning aims to sequentially learn new tasks without forgetting previous tasks' knowledge (catastrophic forgetting). One factor that can cause forgetting is the interference between the gradients on losses from different tasks. When the gradients on the current task's loss are in opposing directions to those on previous tasks' losses, updating the model for the current task may cause performance degradation on previous tasks. In this paper, we first identify causes of the above interference, and hypothesize that correlations between data representations are a key factor of interference. We then propose a method for promoting appropriate correlations between arbitrary tasks' data representations (i.e., global alignment) in individual task learning. Specifically, we learn the data representation as a taskspecific composition of pre-trained token representations shared across all tasks.
Neural Information Processing Systems
May-30-2025, 13:02:00 GMT
- Country:
- North America > United States > Louisiana (0.14)
- Genre:
- Research Report > Experimental Study (0.93)
- Industry:
- Education (0.68)
- Government (0.67)
- Technology: