ScopeQA: A Framework for Generating Out-of-Scope Questions for RAG
Peng, Zhiyuan, Nian, Jinming, Evfimievski, Alexandre, Fang, Yi
–arXiv.org Artificial Intelligence
Conversational AI agents use Retrieval Augmented Generation (RAG) to provide verifiable document-grounded responses to user inquiries. However, many natural questions do not have good answers: about 25\% contain false assumptions~\cite{Yu2023:CREPE}, and over 50\% are ambiguous~\cite{DBLP:conf/emnlp/MinMHZ20}. RAG agents need high-quality data to improve their responses to confusing questions. This paper presents a novel guided hallucination-based method to efficiently generate a diverse set of borderline out-of-scope confusing questions for a given document corpus. We conduct an empirical comparative evaluation of several large language models as RAG agents to measure the accuracy of confusion detection and appropriate response generation. We contribute a benchmark dataset to the public domain.
arXiv.org Artificial Intelligence
Dec-19-2024
- Country:
- Asia (0.68)
- Europe (0.68)
- North America > United States (0.69)
- Genre:
- Research Report (0.64)
- Industry:
- Leisure & Entertainment > Sports > Football (1.00)
- Technology: