GRS-QA -- Graph Reasoning-Structured Question Answering Dataset
Pahilajani, Anish, Trivedi, Devasha, Shuai, Jincen, Yone, Khin S., Jain, Samyak Rajesh, Park, Namyong, Rossi, Ryan A., Ahmed, Nesreen K., Dernoncourt, Franck, Wang, Yu
–arXiv.org Artificial Intelligence
Large Language Models (LLMs) have excelled in multi-hop question-answering (M-QA) due to their advanced reasoning abilities. However, the impact of the inherent reasoning structures on LLM M-QA performance remains unclear, largely due to the absence of QA datasets that provide fine-grained reasoning structures. To address this gap, we introduce the Graph Reasoning-Structured Question Answering Dataset (GRS-QA), which includes both semantic contexts and reasoning structures for QA pairs. Unlike existing M-QA datasets, where different reasoning structures are entangled together, GRS-QA explicitly captures intricate reasoning pathways by constructing reasoning graphs, where nodes represent textual contexts and edges denote logical flows. These reasoning graphs of different structures enable a fine-grained evaluation of LLM reasoning capabilities across various reasoning structures. Our empirical analysis reveals that LLMs perform differently when handling questions with varying reasoning structures. This finding facilitates the exploration of textual structures as compared with semantics.
arXiv.org Artificial Intelligence
Nov-7-2024
- Country:
- Africa > Middle East
- Morocco (0.04)
- Asia
- Europe
- Germany (0.14)
- Middle East > Malta
- Eastern Region > Northern Harbour District > St. Julian's (0.04)
- Russia > Central Federal District
- Moscow Oblast > Moscow (0.04)
- Ukraine (0.04)
- United Kingdom (0.04)
- North America > United States
- Arkansas > White County (0.04)
- California
- Santa Cruz County > Santa Cruz (0.04)
- Stanislaus County (0.14)
- Tuolumne County (0.04)
- Oregon (0.04)
- Africa > Middle East
- Genre:
- Research Report (1.00)
- Industry:
- Leisure & Entertainment > Sports > Tennis (0.46)
- Technology: