On the Inevitability of Left-Leaning Political Bias in Aligned Language Models
–arXiv.org Artificial Intelligence
The guiding principle of AI alignment is to train large language models (LLMs) to be harmless, helpful, and honest (HHH). At the same time, there are mounting concerns that LLMs exhibit a left-wing political bias. Yet, the commitment to AI alignment cannot be harmonized with the latter critique. In this article, I argue that intelligent systems that are trained to be harmless and honest must necessarily exhibit left-wing political bias. Normative assumptions underlying alignment objectives inherently concur with progressive moral frameworks and left-wing principles, emphasizing harm avoidance, inclusivity, fairness, and empirical truthfulness. Conversely, right-wing ideologies often conflict with alignment guidelines. Yet, research on political bias in LLMs is consistently framing its insights about left-leaning tendencies as a risk, as problematic, or concerning. This way, researchers are actively arguing against AI alignment, tacitly fostering the violation of HHH principles.
arXiv.org Artificial Intelligence
Jul-22-2025
- Country:
- Europe
- France (0.04)
- Germany > Baden-Württemberg
- Stuttgart Region > Stuttgart (0.05)
- Spain (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- North America > United States
- Massachusetts > Middlesex County
- Cambridge (0.04)
- New York > New York County
- New York City (0.04)
- Massachusetts > Middlesex County
- South America > Brazil (0.04)
- Europe
- Genre:
- Overview (0.46)
- Questionnaire & Opinion Survey (0.46)
- Research Report (0.40)
- Industry:
- Education (0.93)
- Government (1.00)
- Health & Medicine > Therapeutic Area (0.48)
- Technology: