Large Language Models Require Curated Context for Reliable Political Fact-Checking -- Even with Reasoning and Web Search
DeVerna, Matthew R., Yang, Kai-Cheng, Yan, Harry Yaojun, Menczer, Filippo
–arXiv.org Artificial Intelligence
Large language models (LLMs) have raised hopes for automated end-to-end fact-checking, but prior studies report mixed results. As mainstream chatbots increasingly ship with reasoning capabilities and web search tools -- and millions of users already rely on them for verification -- rigorous evaluation is urgent. We evaluate 15 recent LLMs from OpenAI, Google, Meta, and DeepSeek on more than 6,000 claims fact-checked by PolitiFact, comparing standard models with reasoning- and web-search variants. Standard models perform poorly, reasoning offers minimal benefits, and web search provides only moderate gains, despite fact-checks being available on the web. In contrast, a curated RAG system using PolitiFact summaries improved macro F1 by 233% on average across model variants. These findings suggest that giving models access to curated high-quality context is a promising path for automated fact-checking.
arXiv.org Artificial Intelligence
Nov-25-2025
- Country:
- Asia
- China (0.04)
- North Korea (0.04)
- Europe > Switzerland (0.04)
- North America > United States
- Indiana (0.04)
- New York > Broome County
- Binghamton (0.04)
- Texas (0.04)
- Asia
- Genre:
- Research Report
- Experimental Study (0.48)
- New Finding (0.66)
- Research Report
- Industry:
- Government > Regional Government
- Media > News (1.00)
- Technology: