How Well Do LLMs Predict Human Behavior? A Measure of their Pretrained Knowledge
Gao, Wayne, Han, Sukjin, Liang, Annie
Large language models (LLMs) are increasingly used in economics as predictive tools--both to generate synthetic responses in place of human subjects (Horton, 2023; Anthis et al., 2025), and to forecast economic outcomes directly (Hewitt et al., 2024a; Faria-e Castro and Leibovici, 2024; Chan-Lau et al., 2025). Their appeal in these roles is obvious: A pretrained LLM embeds a vast amount of information and can be deployed at negligible cost, often in settings where collecting new, domain-specific human data would be expensive or infeasible. What remains unclear is how to assess the quality of these predictions. This paper proposes a measure that quantifies the domain-specific value of LLMs in an interpretable unit: the amount of human data they substitute for. Specifically, we ask how much human data would be required for a conventional model trained on that data to match the predictive performance of the pretrained LLM in that domain.
Jan-21-2026
- Country:
- Asia > Middle East
- Israel (0.04)
- North America > United States
- Pennsylvania (0.04)
- Tennessee (0.04)
- Asia > Middle East
- Genre:
- Research Report > Experimental Study (0.93)
- Industry:
- Banking & Finance > Economy (0.67)
- Government > Regional Government
- Health & Medicine (0.93)
- Technology: