Seeing the Unseen: How EMoE Unveils Bias in Text-to-Image Diffusion Models
Berry, Lucas, Brando, Axel, Chang, Wei-Di, Higuera, Juan Camilo Gamboa, Meger, David
–arXiv.org Artificial Intelligence
Estimating uncertainty in text-to-image diffusion models is challenging because of their large parameter counts (often exceeding 100 million) and operation in complex, high-dimensional spaces with virtually infinite input possibilities. In this paper, we propose Epistemic Mixture of Experts (EMoE), a novel framework for efficiently estimating epistemic uncertainty in diffusion models. EMoE leverages pre-trained networks without requiring additional training, enabling direct uncertainty estimation from a prompt. We leverage a latent space within the diffusion process that captures epistemic uncertainty better than existing methods. Experimental results on the COCO dataset demonstrate EMoE's effectiveness, showing a strong correlation between uncertainty and image quality. Additionally, EMoE identifies under-sampled languages and regions with higher uncertainty, revealing hidden biases in the training set. This capability demonstrates the relevance of EMoE as a tool for addressing fairness and accountability in AI-generated content.
arXiv.org Artificial Intelligence
May-20-2025
- Country:
- Asia
- China > Tibet Autonomous Region (0.04)
- Middle East > Jordan (0.04)
- Europe
- Germany > Bavaria
- Upper Bavaria > Munich (0.04)
- Italy > Calabria
- Catanzaro Province > Catanzaro (0.04)
- Switzerland > Zürich
- Zürich (0.14)
- Germany > Bavaria
- North America > Canada
- Asia
- Genre:
- Research Report > New Finding (1.00)
- Technology: