Cognitive bias in large language models: Cautious optimism meets anti-Panglossian meliorism
–arXiv.org Artificial Intelligence
The recent success of large language models gives new urgency to the question of how model performance should be evaluated. In many tasks, models can be evaluated for the accuracy of their outputs. However, models can also be evaluated along other important dimensions. For example, we can assess models for the transparency or interpretability of their judgments (Creel 2020; Vredenburgh 2022). We can also assess models for the presence of problematic biases (Kelly 2023; Johnson 2020). Most work on biases in large language models focuses on a conception of bias closely tied to unfairness, especially as affecting marginalized social groups. However, recent work has alleged that large language models also show a number of classic cognitive biases familiar from work in the psychology of reasoning, behavioral economics, and judgment and decisionmaking (Dasgupta et al. 2022; Lin and Ng 2023; Jones and Steinhardt 2022). This development is exciting because it raises the possibility of using cognitive bias as a novel metric by which to evaluate the performance of large language models.
arXiv.org Artificial Intelligence
Nov-17-2023
- Country:
- Africa (0.04)
- Asia > Middle East
- Jordan (0.04)
- Europe > United Kingdom
- England
- Cambridgeshire > Cambridge (0.04)
- Oxfordshire > Oxford (0.04)
- England
- North America
- Canada > Ontario
- Toronto (0.04)
- Puerto Rico > Peñuelas
- Peñuelas (0.04)
- United States > Georgia
- Fulton County > Atlanta (0.04)
- Canada > Ontario
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Health & Medicine (0.47)