Automated Knowledge Graph Construction using Large Language Models and Sentence Complexity Modelling
Anuyah, Sydney, Kaushik, Mehedi Mahmud, Dwarampudi, Krishna, Shiradkar, Rakesh, Durresi, Arjan, Chakraborty, Sunandan
–arXiv.org Artificial Intelligence
We introduce CoDe-KG, an open-source, end-to-end pipeline for extracting sentence-level knowledge graphs by combining robust coreference resolution with syntactic sentence decomposition. Using our model, we contribute a dataset of over 150,000 knowledge triples, which is open source. We also contribute a training corpus of 7248 rows for sentence complexity, 190 rows of gold human annotations for co-reference resolution using open source lung-cancer abstracts from PubMed, 900 rows of gold human annotations for sentence conversion policies, and 398 triples of gold human annotations. We systematically select optimal prompt-model pairs across five complexity categories, showing that hybrid chain-of-thought and few-shot prompting yields up to 99.8% exact-match accuracy on sentence simplification. On relation extraction (RE), our pipeline achieves 65.8% macro-F1 on REBEL, an 8-point gain over the prior state of the art, and 75.7% micro-F1 on WebNLG2, while matching or exceeding performance on Wiki-NRE and CaRB. Ablation studies demonstrate that integrating coreference and decomposition increases recall on rare relations by over 20%. Code and dataset are available at https://github.com/KaushikMahmud/CoDe-KG_EMNLP_2025
arXiv.org Artificial Intelligence
Nov-13-2025
- Country:
- Africa > Nigeria (0.04)
- Asia > China
- Hong Kong (0.04)
- Europe > United Kingdom
- England > West Yorkshire > Leeds (0.05)
- North America
- Canada (0.04)
- United States > Indiana
- Marion County > Indianapolis (0.04)
- Genre:
- Research Report (1.00)
- Industry:
- Health & Medicine > Therapeutic Area > Oncology (1.00)
- Technology: