Imputation Strategies Under Clinical Presence: Impact on Algorithmic Fairness
Jeanselme, Vincent, De-Arteaga, Maria, Zhang, Zhe, Barrett, Jessica, Tom, Brian
–arXiv.org Artificial Intelligence
Machine learning risks reinforcing biases present in data, and, as we argue in this work, in what is absent from data. In healthcare, biases have marked medical history, leading to unequal care affecting marginalised groups. Patterns in missing data often reflect these group discrepancies, but the algorithmic fairness implications of group-specific missingness are not well understood. Despite its potential impact, imputation is often an overlooked preprocessing step, with attention placed on the reduction of reconstruction error and overall performance, ignoring how imputation can affect groups differently. Our work studies how imputation choices affect reconstruction errors across groups and algorithmic fairness properties of downstream predictions.
arXiv.org Artificial Intelligence
Jun-30-2023
- Country:
- Asia > India (0.04)
- Europe
- Belgium (0.04)
- Netherlands (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.14)
- North America > United States
- California > San Diego County
- San Diego (0.04)
- Georgia > Richmond County
- Augusta (0.04)
- Texas > Travis County
- Austin (0.04)
- Virginia (0.04)
- California > San Diego County
- Oceania > New Zealand (0.04)
- Genre:
- Overview (0.92)
- Research Report
- Experimental Study (1.00)
- New Finding (1.00)
- Industry:
- Health & Medicine
- Diagnostic Medicine (0.67)
- Health Care Access (0.67)
- Health Care Technology (0.68)
- Therapeutic Area
- Cardiology/Vascular Diseases (1.00)
- Oncology (0.93)
- Health & Medicine
- Technology: