Capturing Symmetry and Antisymmetry in Language Models through Symmetry-Aware Training Objectives
Yuan, Zhangdie, Vlachos, Andreas
–arXiv.org Artificial Intelligence
Capturing symmetric (e.g., country borders another country) and antisymmetric (e.g., parent_of) relations is crucial for a variety of applications. This paper tackles this challenge by introducing a novel Wikidata-derived natural language inference dataset designed to evaluate large language models (LLMs). Our findings reveal that LLMs perform comparably to random chance on this benchmark, highlighting a gap in relational understanding. To address this, we explore encoder retraining via contrastive learning with k-nearest neighbors. The retrained encoder matches the performance of fine-tuned classification heads while offering additional benefits, including greater efficiency in few-shot learning and improved mitigation of catastrophic forgetting.
arXiv.org Artificial Intelligence
Apr-24-2025
- Country:
- Asia > Singapore (0.05)
- Europe
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Ireland > Leinster
- Genre:
- Research Report > New Finding (1.00)