Making Language Models Robust Against Negation
Rezaei, MohammadHossein, Blanco, Eduardo
–arXiv.org Artificial Intelligence
Negation has been a long-standing challenge for language models. Previous studies have shown that they struggle with negation in many natural language understanding tasks. In this work, we propose a self-supervised method to make language models more robust against negation. We introduce a novel task, Next Sentence Polarity Prediction (NSPP), and a variation of the Next Sentence Prediction (NSP) task. We show that BERT and RoBERTa further pre-trained on our tasks outperform the off-the-shelf versions on nine negation-related benchmarks. Most notably, our pre-training tasks yield between 1.8% and 9.1% improvement on CondaQA, a large question-answering corpus requiring reasoning over negation.
arXiv.org Artificial Intelligence
Feb-11-2025
- Country:
- Asia (1.00)
- Europe (1.00)
- North America > United States
- Maryland (0.28)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- Genre:
- Research Report (0.82)
- Technology: