CorefInst: Leveraging LLMs for Multilingual Coreference Resolution
Arslan, Tuğba Pamay, Erol, Emircan, Eryiğit, Gülşen
–arXiv.org Artificial Intelligence
Coreference Resolution (CR) is a crucial yet challenging task in natural language understanding, often constrained by task-specific architectures and encoder-based language models that demand extensive training and lack adaptability. This study introduces the first multilingual CR methodology which leverages decoder-only LLMs to handle both overt and zero mentions. The article explores how to model the CR task for LLMs via five different instruction sets using a controlled inference method. The approach is evaluated across three LLMs; Llama 3.1, Gemma 2, and Mistral 0.3. The results indicate that LLMs, when instruction-tuned with a suitable instruction set, can surpass state-of-the-art task-specific architectures. Specifically, our best model, a fully fine-tuned Llama 3.1 for multilingual CR, outperforms the leading multilingual CR model (i.e., Corpipe 24 single stage variant) by 2 pp on average across all languages in the CorefUD v1.2 dataset collection.
arXiv.org Artificial Intelligence
Sep-23-2025
- Country:
- Asia
- Middle East > Republic of Türkiye
- Istanbul Province > Istanbul (0.04)
- Singapore (0.04)
- Middle East > Republic of Türkiye
- Europe
- Germany
- Berlin (0.04)
- Brandenburg > Potsdam (0.04)
- Hungary > Csongrád-Csanád County
- Szeged (0.05)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Middle East > Republic of Türkiye
- Istanbul Province > Istanbul (0.04)
- Germany
- Asia
- Genre:
- Research Report
- Experimental Study (0.68)
- New Finding (1.00)
- Research Report
- Technology: