HRET: A Self-Evolving LLM Evaluation Toolkit for Korean
Lee, Hanwool, Kim, Soo Yong, Choi, Dasol, Baek, SangWon, Hong, Seunghyeok, Jeong, Ilgyun, Hwang, Inseon, Lee, Naeun, Son, Guijin
–arXiv.org Artificial Intelligence
Recent advancements in Korean large language models (LLMs) have spurred numerous benchmarks and evaluation methodologies, yet the lack of a standardized evaluation framework has led to inconsistent results and limited comparability. To address this, we introduce HRET Haerae Evaluation Toolkit, an open-source, self-evolving evaluation framework tailored specifically for Korean LLMs. HRET unifies diverse evaluation methods, including logit-based scoring, exact-match, language-inconsistency penalization, and LLM-as-a-Judge assessments. Its modular, registry-based architecture integrates major benchmarks (HAE-RAE Bench, KMMLU, KUDGE, HRM8K) and multiple inference backends (vLLM, HuggingFace, OpenAI-compatible endpoints). With automated pipelines for continuous evolution, HRET provides a robust foundation for reproducible, fair, and transparent Korean NLP research.
arXiv.org Artificial Intelligence
Apr-1-2025