MaXIFE: Multilingual and Cross-lingual Instruction Following Evaluation
Liu, Yile, Ma, Ziwei, Jiang, Xiu, Hu, Jinglu, Chang, Jing, Li, Liang
–arXiv.org Artificial Intelligence
With the rapid adoption of large language models (LLMs) in natural language processing, the ability to follow instructions has emerged as a key metric for evaluating their practical utility. However, existing evaluation methods often focus on single-language scenarios, overlooking the challenges and differences present in multilingual and cross-lingual contexts. To address this gap, we introduce MaXIFE: a comprehensive evaluation benchmark designed to assess instruction-following capabilities across 23 different languages with 1667 verifiable instruction tasks. MaXIFE integrates both Rule-Based Evaluation and Model-Based Evaluation, ensuring a balance of efficiency and accuracy. We applied MaXIFE to evaluate several leading commercial LLMs, establishing baseline results for future comparisons. By providing a standardized tool for multilingual instruction-following evaluation, MaXIFE aims to advance research and development in natural language processing.
arXiv.org Artificial Intelligence
Jun-4-2025
- Country:
- Africa > Niger (0.04)
- Asia
- China > Guangdong Province
- Shenzhen (0.04)
- Middle East > Iraq
- Kurdistan Region > Duhok Governorate > Zakho (0.04)
- Thailand > Bangkok
- Bangkok (0.04)
- China > Guangdong Province
- Europe > Ireland
- Leinster > County Dublin > Dublin (0.04)
- North America
- Dominican Republic (0.04)
- Mexico > Mexico City
- Mexico City (0.04)
- United States (0.04)
- South America > Peru
- Cusco Department > Cusco Province > Cusco (0.04)
- Genre:
- Research Report (0.81)
- Industry:
- Education (0.67)
- Information Technology > Security & Privacy (0.45)
- Leisure & Entertainment (0.45)
- Technology: