Attesting Biases and Discrimination using Language Semantics
Aran, Xavier Ferrer, Such, Jose M., Criado, Natalia
–arXiv.org Artificial Intelligence
AI agents are increasingly deployed and used to make automated decisions that affect our lives on a daily basis. It is imperative to ensure that these systems embed ethical principles and respect human values. We focus on how we can attest to whether AI agents treat users fairly without discriminating against particular individuals or groups through biases in language. In particular, we discuss human unconscious biases, how they are embedded in language, and how AI systems inherit those biases by learning from and processing human language. Then, we outline a roadmap for future research to better understand and attest problematic AI biases derived from language.
arXiv.org Artificial Intelligence
Sep-10-2019
- Country:
- Asia > China
- Hong Kong (0.04)
- Europe
- Belgium (0.04)
- France (0.04)
- United Kingdom
- England
- Greater London > London (0.04)
- Oxfordshire > Oxford (0.04)
- Wales (0.04)
- England
- North America > United States (0.14)
- Asia > China
- Genre:
- Research Report (0.64)
- Industry:
- Information Technology (0.46)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning (1.00)
- Natural Language > Chatbot (0.94)
- Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence