BUSTED at AraGenEval Shared Task: A Comparative Study of Transformer-Based Models for Arabic AI-Generated Text Detection
Zain, Ali, Farooqui, Sareem, Rafi, Muhammad
–arXiv.org Artificial Intelligence
This paper details our submission to the AraGenEval Shared Task on Arabic AI-generated text detection, where our team, BUSTED, secured 5th place. We investigated the effectiveness of three pre-trained transformer models: AraELECTRA, CAMeLBERT, and XLM-RoBERTa. Our approach involved fine-tuning each model on the provided dataset for a binary classification task. Our findings revealed a surprising result: the multilingual XLM-RoBERTa model achieved the highest performance with an F1 score of 0.7701, outperforming the specialized Arabic models. This work underscores the complexities of AI-generated text detection and highlights the strong generalization capabilities of multilingual models.
arXiv.org Artificial Intelligence
Oct-28-2025
- Country:
- Asia
- China (0.04)
- Middle East
- Israel (0.04)
- Palestine > Gaza Strip
- Gaza Governorate > Gaza (0.04)
- Pakistan > Sindh
- Karachi Division > Karachi (0.05)
- North America > United States (0.04)
- Asia
- Genre:
- Research Report > New Finding (0.67)
- Technology: