Owls are wise and foxes are unfaithful: Uncovering animal stereotypes in vision-language models
Aman, Tabinda, Nadeem, Mohammad, Sohail, Shahab Saquib, Anas, Mohammad, Cambria, Erik
–arXiv.org Artificial Intelligence
Generative artificial intelligence (GAI) has seen rapid adoption across diverse domains through its ability to produce high-quality text, images, and videos [1]. Vision-Language Models (VLMs) represent a significant advancement in this space, combining visual and linguistic understanding to generate contextually relevant images from textual descriptions [2]. They leverage vast datasets and sophisticated algorithms [2,3] to enable unprecedented creativity and efficiency, driving applications in marketing, entertainment, design, and more. Large Language Models (LLMs) and VLMs often inherit and perpetuate biases and stereotypes present in their training data [4-7], which is typically sourced from vast and diverse internet repositories [8-11]. The training datasets frequently contain implicit and explicit cultural stereotypes, societal biases, and skewed representations that the models learn during training.
arXiv.org Artificial Intelligence
Jan-21-2025
- Country:
- Asia
- India
- Madhya Pradesh > Bhopal (0.04)
- NCT > New Delhi (0.04)
- Uttar Pradesh > Aligarh (0.05)
- Singapore (0.04)
- India
- Asia
- Genre:
- Research Report (0.65)
- Technology: