FairJudge: MLLM Judging for Social Attributes and Prompt Image Alignment
Sahili, Zahraa Al, Fetanat, Maryam, Nowaz, Maimuna, Patras, Ioannis, Purver, Matthew
–arXiv.org Artificial Intelligence
Text-to-image (T2I) systems lack simple, reproducible ways to evaluate how well images match prompts and how models treat social attributes. Common proxies -- face classifiers and contrastive similarity -- reward surface cues, lack calibrated abstention, and miss attributes only weakly visible (for example, religion, culture, disability). We present FairJudge, a lightweight protocol that treats instruction-following multimodal LLMs as fair judges. It scores alignment with an explanation-oriented rubric mapped to [-1, 1]; constrains judgments to a closed label set; requires evidence grounded in the visible content; and mandates abstention when cues are insufficient. Unlike CLIP-only pipelines, FairJudge yields accountable, evidence-aware decisions; unlike mitigation that alters generators, it targets evaluation fairness. We evaluate gender, race, and age on FairFace, PaTA, and FairCoT; extend to religion, culture, and disability; and assess profession correctness and alignment on IdenProf, FairCoT-Professions, and our new DIVERSIFY-Professions. We also release DIVERSIFY, a 469-image corpus of diverse, non-iconic scenes. Across datasets, judge models outperform contrastive and face-centric baselines on demographic prediction and improve mean alignment while maintaining high profession accuracy, enabling more reliable, reproducible fairness audits.
arXiv.org Artificial Intelligence
Nov-20-2025
- Country:
- Europe
- Slovenia > Central Slovenia
- Municipality of Ljubljana > Ljubljana (0.04)
- United Kingdom > England
- Greater London > London (0.04)
- Slovenia > Central Slovenia
- North America > United States
- Florida > Miami-Dade County > Miami (0.04)
- Europe
- Genre:
- Research Report (0.82)
- Industry:
- Health & Medicine > Therapeutic Area (0.93)
- Technology: