Adapting Small Language Models to Low-Resource Domains: A Case Study in Hindi Tourism QA
Majhi, Sandipan, Bhattacharya, Paheli
–arXiv.org Artificial Intelligence
Domain-specific question answering in low-resource languages faces two key challenges: scarcity of annotated datasets and limited domain knowledge in general-purpose language models. In this work, we present a multi-stage finetuning strategy to adapt lightweight language models to the Hindi tourism domain by leveraging both original and synthetic training data. Synthetic question-answer pairs are generated using large LLMs (LLaMA-70B, Phi-14B) and used to augment the limited original dataset. We explore several training methodologies and analyse their impact on domain generalisation. Our results demonstrate that large models can efficiently generate synthetic data, while small models can effectively adapt to it, offering a scalable pathway for low-resource, domain-specific QA.
arXiv.org Artificial Intelligence
Oct-30-2025
- Country:
- Asia
- India
- Karnataka > Bengaluru (0.05)
- West Bengal > Kharagpur (0.04)
- Middle East > UAE
- Abu Dhabi Emirate > Abu Dhabi (0.14)
- Thailand > Bangkok
- Bangkok (0.04)
- India
- North America
- Canada > Ontario
- Toronto (0.05)
- Mexico > Mexico City
- Mexico City (0.04)
- United States
- Florida > Miami-Dade County
- Miami (0.04)
- New Mexico > Bernalillo County
- Albuquerque (0.04)
- Florida > Miami-Dade County
- Canada > Ontario
- Asia
- Genre:
- Research Report > New Finding (0.54)
- Industry:
- Consumer Products & Services > Travel (0.62)
- Technology: