Evaluating CxG Generalisation in LLMs via Construction-Based NLI Fine Tuning
Mackintosh, Tom, Madabushi, Harish Tayyar, Bonial, Claire
–arXiv.org Artificial Intelligence
We probe large language models' ability to learn deep form-meaning mappings as defined by construction grammars. We introduce the ConTest-NLI benchmark of 80k sentences covering eight English constructions from highly lexicalized to highly schematic. Our pipeline generates diverse synthetic NLI triples via templating and the application of a model-in-the-loop filter. This provides aspects of human validation to ensure challenge and label reliability. Zero-shot tests on leading LLMs reveal a 24% drop in accuracy between naturalistic (88%) and adversarial data (64%), with schematic patterns proving hardest. Fine-tuning on a subset of ConTest-NLI yields up to 9% improvement, yet our results highlight persistent abstraction gaps in current LLMs and offer a scalable framework for evaluating construction-informed learning.
arXiv.org Artificial Intelligence
Sep-23-2025
- Country:
- Europe > United Kingdom
- England
- Oxfordshire > Oxford (0.04)
- Somerset > Bath (0.04)
- England
- North America > United States
- California > Alameda County
- Berkeley (0.04)
- Illinois > Cook County
- Chicago (0.04)
- California > Alameda County
- Europe > United Kingdom
- Genre:
- Research Report > New Finding (1.00)
- Technology: