Comparative Study of Pre-Trained BERT and Large Language Models for Code-Mixed Named Entity Recognition
Shirke, Mayur, Shembade, Amey, Thorat, Pavan, Wagh, Madhushri, Joshi, Raviraj
–arXiv.org Artificial Intelligence
Named Entity Recognition (NER) in code-mixed text, particularly Hindi-English (Hinglish), presents unique challenges due to informal structure, transliteration, and frequent language switching. This study conducts a comparative evaluation of code-mixed fine-tuned models and non-code-mixed multilingual models, along with zero-shot generative large language models (LLMs). Specifically, we evaluate HingBERT, HingM-BERT, and HingRoBERTa (trained on code-mixed data), and BERT Base Cased, IndicBERT, RoBERTa and MuRIL (trained on non-code-mixed multilingual data). We also assess the performance of Google Gemini in a zero-shot setting using a modified version of the dataset with NER tags removed. All models are tested on a benchmark Hinglish NER dataset using Precision, Recall, and F1-score. Results show that code-mixed models, particularly HingRoBERTa and HingBERT -based fine-tuned models, outperform others -- including closed-source LLMs like Google Gemini -- due to domain-specific pretraining. Non-code-mixed models perform reasonably but show limited adaptability. Notably, Google Gemini exhibits competitive zero-shot performance, underlining the generalization strength of modern LLMs. This study provides key insights into the effectiveness of specialized versus generalized models for code-mixed NER tasks.
arXiv.org Artificial Intelligence
Sep-3-2025
- Country:
- Asia
- India
- Maharashtra > Pune (0.14)
- Tamil Nadu > Chennai (0.04)
- Indonesia > Bali (0.05)
- India
- Asia
- Genre:
- Research Report > New Finding (0.89)
- Technology: