Aligned but Stereotypical? The Hidden Influence of System Prompts on Social Bias in LVLM-Based Text-to-Image Models
Park, NaHyeon, An, Namin, Kim, Kunhee, Yoon, Soyeon, Huo, Jiahao, Shim, Hyunjung
–arXiv.org Artificial Intelligence
Large vision-language model (LVLM) based text-to-image (T2I) systems have become the dominant paradigm in image generation, yet whether they amplify social biases remains insufficiently understood. In this paper, we show that LVLM-based models produce markedly more socially biased images than non-LVLM-based models. We introduce a 1,024 prompt benchmark spanning four levels of linguistic complexity and evaluate demographic bias across multiple attributes in a systematic manner. Our analysis identifies system prompts, the predefined instructions guiding LVLMs, as a primary driver of biased behavior. Through decoded intermediate representations, token-probability diagnostics, and embedding-association analyses, we reveal how system prompts encode demographic priors that propagate into image synthesis. To this end, we propose FairPro, a training-free meta-prompting framework that enables LVLMs to self-audit and construct fairness-aware system prompts at test time. Experiments on two LVLM-based T2I models, SANA and Qwen-Image, show that FairPro substantially reduces demographic bias while preserving text-image alignment. We believe our findings provide deeper insight into the central role of system prompts in bias propagation and offer a practical, deployable approach for building more socially responsible T2I systems.
arXiv.org Artificial Intelligence
Dec-5-2025
- Country:
- Asia > China > Tibet Autonomous Region (0.04)
- Genre:
- Research Report > New Finding (0.48)
- Industry:
- Health & Medicine (1.00)
- Transportation
- Air (0.68)
- Infrastructure & Services (0.46)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning > Neural Networks (0.89)
- Natural Language > Text Processing (0.93)
- Vision (1.00)
- Information Technology > Artificial Intelligence