Leveraging In-Context Learning for Political Bias Testing of LLMs
Haller, Patrick, Vamvas, Jannis, Sennrich, Rico, Jäger, Lena A.
–arXiv.org Artificial Intelligence
A growing body of work has been querying LLMs with political questions to evaluate their potential biases. However, this probing method has limited stability, making comparisons between models unreliable. In this paper, we argue that LLMs need more context. We propose a new probing task, Questionnaire Modeling (QM), that uses human survey data as in-context examples. We show that QM improves the stability of question-based bias evaluation, and demonstrate that it may be used to compare instruction-tuned models to their base versions. Experiments with LLMs of various sizes indicate that instruction tuning can indeed change the direction of bias. Furthermore, we observe a trend that larger models are able to leverage in-context examples more effectively, and generally exhibit smaller bias scores in QM. Data and code are publicly available.
arXiv.org Artificial Intelligence
Jun-30-2025
- Country:
- Asia
- Middle East > UAE
- Abu Dhabi Emirate > Abu Dhabi (0.04)
- Thailand > Bangkok
- Bangkok (0.04)
- Middle East > UAE
- Europe
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Switzerland > Zürich
- Zürich (0.04)
- Ukraine (0.04)
- Belgium > Brussels-Capital Region
- North America
- Canada > Ontario
- Toronto (0.04)
- Mexico > Mexico City
- Mexico City (0.04)
- United States > Florida
- Miami-Dade County > Miami (0.04)
- Canada > Ontario
- Asia
- Genre:
- Questionnaire & Opinion Survey (1.00)
- Research Report > New Finding (0.68)
- Industry:
- Banking & Finance (1.00)
- Energy > Power Industry (0.67)
- Government
- Foreign Policy (0.46)
- Voting & Elections (0.46)
- Health & Medicine (1.00)
- Law (1.00)
- Technology: