Investigating Recent Large Language Models for Vietnamese Machine Reading Comprehension
Nguyen, Anh Duc, Phi, Hieu Minh, Ngo, Anh Viet, Trieu, Long Hai, Nguyen, Thai Phuong
–arXiv.org Artificial Intelligence
Large Language Models (LLMs) have shown remarkable proficiency in Machine Reading Comprehension (MRC) tasks; however, their effectiveness for low-resource languages like Vietnamese remains largely unexplored. In this paper, we fine-tune and evaluate two state-of-the-art LLMs: Llama 3 (8B parameters) and Gemma (7B parameters), on ViMMRC, a Vietnamese MRC dataset. By utilizing Quantized Low-Rank Adaptation (QLoRA), we efficiently fine-tune these models and compare their performance against powerful LLM-based baselines. Although our fine-tuned models are smaller than GPT-3 and GPT-3.5, they outperform both traditional BERT-based approaches and these larger models. This demonstrates the effectiveness of our fine-tuning process, showcasing how modern LLMs can surpass the capabilities of older models like BERT while still being suitable for deployment in resource-constrained environments. Through intensive analyses, we explore various aspects of model performance, providing valuable insights into adapting LLMs for low-resource languages like Vietnamese. Our study contributes to the advancement of natural language processing in low-resource languages, and we make our fine-tuned models publicly available at: https://huggingface.co/iaiuet.
arXiv.org Artificial Intelligence
Mar-23-2025
- Country:
- North America > United States > California (0.28)
- Genre:
- Research Report > New Finding (0.68)
- Industry:
- Education > Assessment & Standards > Student Performance (0.62)
- Technology: