Ustnlp16 at SemEval-2025 Task 9: Improving Model Performance through Imbalance Handling and Focal Loss
Cai, Zhuoang, Li, Zhenghao, Liu, Yang, Guo, Liyuan, Song, Yangqiu
–arXiv.org Artificial Intelligence
Classification tasks often suffer from imbal- anced data distribution, which presents chal- lenges in food hazard detection due to severe class imbalances, short and unstructured text, and overlapping semantic categories. In this paper, we present our system for SemEval- 2025 Task 9: Food Hazard Detection, which ad- dresses these issues by applying data augmenta- tion techniques to improve classification perfor- mance. We utilize transformer-based models, BERT and RoBERTa, as backbone classifiers and explore various data balancing strategies, including random oversampling, Easy Data Augmentation (EDA), and focal loss. Our ex- periments show that EDA effectively mitigates class imbalance, leading to significant improve- ments in accuracy and F1 scores. Furthermore, combining focal loss with oversampling and EDA further enhances model robustness, par- ticularly for hard-to-classify examples. These findings contribute to the development of more effective NLP-based classification models for food hazard detection.
arXiv.org Artificial Intelligence
May-2-2025
- Country:
- Asia > China
- Hong Kong (0.05)
- Europe > Austria
- Vienna (0.14)
- North America > United States
- Florida > Miami-Dade County > Miami (0.04)
- Asia > China
- Genre:
- Research Report > New Finding (0.95)
- Technology: