Are Hallucinations Bad Estimations?
Liu, Hude, Hu, Jerry Yao-Chieh, Zhang, Jennifer Yuntong, Song, Zhao, Liu, Han
We formalize hallucinations in generative models as failures to link an estimate to any plausible cause. Under this interpretation, we show that even loss-minimizing optimal estimators still hallucinate. We confirm this with a general high probability lower bound on hallucinate rate for generic data distributions. This reframes hallucination as structural misalignment between loss minimization and human-acceptable outputs, and hence estimation errors induced by miscalibration. Experiments on coin aggregation, open-ended QA, and text-to-image support our theory.
Sep-29-2025
- Country:
- North America
- Canada > Ontario
- Toronto (0.14)
- United States
- California
- Alameda County > Berkeley (0.14)
- San Francisco County > San Francisco (0.14)
- Illinois > Cook County
- California
- Canada > Ontario
- North America
- Genre:
- Research Report > New Finding (0.46)
- Technology: