Hypothetical Documents or Knowledge Leakage? Rethinking LLM-based Query Expansion
Yoon, Yejun, Jung, Jaeyoon, Yoon, Seunghyun, Park, Kunwoo
–arXiv.org Artificial Intelligence
Query expansion methods powered by large language models (LLMs) have demonstrated effectiveness in zero-shot retrieval tasks. These methods assume that LLMs can generate hypothetical documents that, when incorporated into a query vector, enhance the retrieval of real evidence. However, we challenge this assumption by investigating whether knowledge leakage in benchmarks contributes to the observed performance gains. Using fact verification as a testbed, we analyze whether the generated documents contain information entailed by ground-truth evidence and assess their impact on performance. Our findings indicate that, on average, performance improvements consistently occurred for claims whose generated documents included sentences entailed by gold evidence. This suggests that knowledge leakage may be present in fact-verification benchmarks, potentially inflating the perceived performance of LLM-based query expansion methods.
arXiv.org Artificial Intelligence
Jun-5-2025
- Country:
- Africa > Nigeria (0.04)
- Asia
- Middle East > UAE
- Abu Dhabi Emirate > Abu Dhabi (0.14)
- Singapore (0.04)
- Thailand > Bangkok
- Bangkok (0.04)
- Middle East > UAE
- Europe > Middle East
- Malta > Eastern Region > Northern Harbour District > St. Julian's (0.04)
- North America
- Canada > Ontario
- Toronto (0.04)
- Mexico > Mexico City
- Mexico City (0.04)
- United States
- Florida > Miami-Dade County
- Miami (0.04)
- Louisiana > Orleans Parish
- New Orleans (0.04)
- Michigan > Washtenaw County
- Ann Arbor (0.04)
- Nebraska > Hall County
- Grand Island (0.14)
- Florida > Miami-Dade County
- Canada > Ontario
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Technology: