How Aligned are Generative Models to Humans in High-Stakes Decision-Making?
Tan, Sarah, Mallari, Keri, Adebayo, Julius, Gordo, Albert, Wells, Martin T., Inkpen, Kori
–arXiv.org Artificial Intelligence
Large generative models (LMs) are increasingly being considered for high-stakes decision-making. This work considers how such models compare to humans and predictive AI models on a specific case of recidivism prediction. We combine three datasets -- COMPAS predictive AI risk scores, human recidivism judgements, and photos -- into a dataset on which we study the properties of several state-of-the-art, multimodal LMs. Beyond accuracy and bias, we focus on studying human-LM alignment on the task of recidivism prediction. We investigate if these models can be steered towards human decisions, the impact of adding photos, and whether anti-discimination prompting is effective. We find that LMs can be steered to outperform humans and COMPAS using in context-learning. We find anti-discrimination prompting to have unintended effects, causing some models to inhibit themselves and significantly reduce their number of positive predictions.
arXiv.org Artificial Intelligence
Oct-20-2024
- Country:
- North America > United States (0.68)
- Genre:
- Research Report
- Experimental Study (0.67)
- New Finding (0.67)
- Research Report
- Industry:
- Government (0.92)
- Health & Medicine (0.93)
- Law > Criminal Law (1.00)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (1.00)
- Technology: