A Distributionally Robust Approach to Fair Classification
Taskesen, Bahar, Nguyen, Viet Anh, Kuhn, Daniel, Blanchet, Jose
We propose a distributionally robust logistic regression model with an unfairness penalty that prevents discrimination with respect to sensitive attributes such as gender or ethnicity. This model is equivalent to a tractable convex optimization problem if a Wasserstein ball centered at the empirical distribution on the training data is used to model distributional uncertainty and if a new convex unfairness measure is used to incentivize equalized opportunities. We demonstrate that the resulting classifier improves fairness at a marginal loss of predictive accuracy on both synthetic and real datasets. We also derive linear programming-based confidence bounds on the level of unfairness of any pre-trained classifier by leveraging techniques from optimal uncertainty quantification over Wasserstein balls.
Jul-18-2020
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe > Switzerland (0.04)
- North America > United States
- California > Santa Clara County > Palo Alto (0.04)
- Asia > Middle East
- Genre:
- Research Report > Experimental Study (0.35)
- Industry:
- Information Technology > Security & Privacy (0.46)
- Law (1.00)
- Technology: