Rationale-Enhanced Language Models are Better Continual Relation Learners
Xiong, Weimin, Song, Yifan, Wang, Peiyi, Li, Sujian
–arXiv.org Artificial Intelligence
Continual relation extraction (CRE) aims to solve the problem of catastrophic forgetting when learning a sequence of newly emerging relations. Recent CRE studies have found that catastrophic forgetting arises from the model's lack of robustness against future analogous relations. To address the issue, we introduce rationale, i.e., the explanations of relation classification results generated by large language models (LLM), into CRE task. Specifically, we design the multi-task rationale tuning strategy to help the model learn current relations robustly. We also conduct contrastive rationale replay to further distinguish analogous relations. Experimental results on two standard benchmarks demonstrate that our method outperforms the state-of-the-art CRE models.
arXiv.org Artificial Intelligence
Oct-10-2023
- Country:
- Asia
- China (0.04)
- Middle East > UAE
- Abu Dhabi Emirate > Abu Dhabi (0.04)
- Russia (0.05)
- Europe
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- Denmark > Capital Region
- Copenhagen (0.04)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Russia (0.05)
- Belgium > Brussels-Capital Region
- North America > United States
- Indiana > Pike County
- Petersburg (0.04)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- Washington > King County
- Seattle (0.04)
- Indiana > Pike County
- Asia
- Genre:
- Research Report > New Finding (0.46)
- Technology: