SimpleQA Verified: A Reliable Factuality Benchmark to Measure Parametric Knowledge
Haas, Lukas, Yona, Gal, D'Antonio, Giovanni, Goldshtein, Sasha, Das, Dipanjan
–arXiv.org Artificial Intelligence
We introduce SimpleQA Verified, a 1,000-prompt benchmark for evaluating Large Language Model (LLM) short-form factuality based on OpenAI's SimpleQA. It addresses critical limitations in OpenAI's benchmark, including noisy and incorrect labels, topical biases, and question redundancy. SimpleQA Verified was created through a rigorous multi-stage filtering process involving de-duplication, topic balancing, and source reconciliation to produce a more reliable and challenging evaluation set, alongside improvements in the autorater prompt. On this new benchmark, Gemini 2.5 Pro achieves a state-of-the-art F1-score of 55.6, outperforming other frontier models, including GPT-5. This work provides the research community with a higher-fidelity tool to track genuine progress in parametric model factuality and to mitigate hallucinations. The benchmark dataset, evaluation code, and leaderboard are available at: https://www.kaggle.com/benchmarks/deepmind/simpleqa-verified.
arXiv.org Artificial Intelligence
Sep-10-2025
- Country:
- Africa
- Malawi > Central Region
- Lilongwe District > Lilongwe (0.04)
- South Africa > Western Cape
- Cape Town (0.04)
- Malawi > Central Region
- Asia
- North America > United States
- California
- Riverside County > Norco (0.04)
- San Francisco County > San Francisco (0.14)
- Massachusetts > Hampshire County
- Amherst (0.04)
- California
- South America > Colombia (0.04)
- Africa
- Genre:
- Research Report (0.64)
- Industry:
- Government (0.69)
- Leisure & Entertainment (1.00)
- Media > Television (0.46)
- Technology: