Bench4KE: Benchmarking Automated Competency Question Generation
Lippolis, Anna Sofia, Ragagni, Minh Davide, Ciancarini, Paolo, Nuzzolese, Andrea Giovanni, Presutti, Valentina
–arXiv.org Artificial Intelligence
The availability of Large Language Models (LLMs) presents a unique opportunity to reinvigorate research on Knowledge Engineering (KE) automation. This trend is already evident in recent efforts developing LLM-based methods and tools for the automatic generation of Competency Questions (CQs), natural language questions used by ontology engineers to define the functional requirements of an ontology. However, the evaluation of these tools lacks standardization. This undermines the methodological rigor and hinders the replication and comparison of results. To address this gap, we introduce Bench4KE, an extensible API-based benchmarking system for KE automation. The presented release focuses on evaluating tools that generate CQs automatically. Bench4KE provides a curated gold standard consisting of CQ datasets from 17 real-world ontology engineering projects and uses a suite of similarity metrics to assess the quality of the CQs generated. We present a comparative analysis of 6 recent CQ generation systems, which are based on LLMs, establishing a baseline for future research. Bench4KE is also designed to accommodate additional KE automation tasks, such as SPARQL query generation, ontology testing and drafting. Code and datasets are publicly available under the Apache 2.0 license.
arXiv.org Artificial Intelligence
Dec-10-2025
- Country:
- Europe
- Austria > Vienna (0.14)
- Ireland (0.04)
- Italy
- Campania > Naples (0.04)
- Emilia-Romagna > Metropolitan City of Bologna
- Bologna (0.05)
- Tuscany > Pisa Province
- Pisa (0.04)
- Switzerland (0.04)
- North America
- Puerto Rico > Peñuelas
- Peñuelas (0.04)
- United States > California
- Monterey County > Monterey (0.04)
- Puerto Rico > Peñuelas
- Oceania > New Zealand
- North Island > Auckland Region > Auckland (0.04)
- Europe
- Genre:
- Research Report (1.00)
- Technology: