Protecting the Protected Group: Circumventing Harmful Fairness
DOI:
https://doi.org/10.1609/aaai.v35i6.16654Keywords:
Game Theory, Ethics -- Bias, Fairness, Transparency & PrivacyAbstract
The recent literature on fair Machine Learning manifests that the choice of fairness constraints must be driven by the utilities of the population. However, virtually all previous work makes the unrealistic assumption that the exact underlying utilities of the population (representing private tastes of individuals) are known to the regulator that imposes the fairness constraint. In this paper we initiate the discussion of the \emph{mismatch}, the unavoidable difference between the underlying utilities of the population and the utilities assumed by the regulator. We demonstrate that the mismatch can make the disadvantaged protected group worse off after imposing the fairness constraint and provide tools to design fairness constraints that help the disadvantaged group despite the mismatch.Downloads
Published
2021-05-18
How to Cite
Ben-Porat, O., Sandomirskiy, F., & Tennenholtz, M. (2021). Protecting the Protected Group: Circumventing Harmful Fairness. Proceedings of the AAAI Conference on Artificial Intelligence, 35(6), 5176-5184. https://doi.org/10.1609/aaai.v35i6.16654
Issue
Section
AAAI Technical Track on Game Theory and Economic Paradigms