On Fairness and Calibration
Geoff Pleiss, Manish Raghavan, Felix Wu, Jon Kleinberg, Kilian Q. Weinberger
–Neural Information Processing Systems
The machine learning community has become increasingly concerned with the potential for bias and discrimination in predictive models. This has motivated a growing line of work on what it means for a classification procedure to be "fair." In this paper, we investigate the tension between minimizing error disparity across different population groups while maintaining calibrated probability estimates. We show that calibration is compatible only with a single error constraint (i.e.
Neural Information Processing Systems
Oct-8-2024, 07:37:22 GMT
- Country:
- North America > United States (0.93)
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Law > Civil Rights & Constitutional Law (0.46)
- Technology: