Arbitrariness and Social Prediction: The Confounding Role of Variance in Fair Classification

Authors

  • A. Feder Cooper The Center for Generative AI, Law, and Policy Research Cornell University
  • Katherine Lee The Center for Generative AI, Law, and Policy Research Cornell University
  • Madiha Zahrah Choksi Cornell University
  • Solon Barocas Cornell University Microsoft Research
  • Christopher De Sa Cornell University
  • James Grimmelmann The Center for Generative AI, Law, and Policy Research Cornell University
  • Jon Kleinberg Cornell University
  • Siddhartha Sen Microsoft Research
  • Baobao Zhang Syracuse University

DOI:

https://doi.org/10.1609/aaai.v38i20.30203

Keywords:

General

Abstract

Variance in predictions across different trained models is a significant, under-explored source of error in fair binary classification. In practice, the variance on some data examples is so large that decisions can be effectively arbitrary. To investigate this problem, we take an experimental approach and make four overarching contributions. We: 1) Define a metric called self-consistency, derived from variance, which we use as a proxy for measuring and reducing arbitrariness; 2) Develop an ensembling algorithm that abstains from classification when a prediction would be arbitrary; 3) Conduct the largest to-date empirical study of the role of variance (vis-a-vis self-consistency and arbitrariness) in fair binary classification; and, 4) Release a toolkit that makes the US Home Mortgage Disclosure Act (HMDA) datasets easily usable for future research. Altogether, our experiments reveal shocking insights about the reliability of conclusions on benchmark datasets. Most fair binary classification benchmarks are close-to-fair when taking into account the amount of arbitrariness present in predictions -- before we even try to apply any fairness interventions. This finding calls into question the practical utility of common algorithmic fairness methods, and in turn suggests that we should reconsider how we choose to measure fairness in binary classification.

Published

2024-03-24

How to Cite

Cooper, A. F., Lee, K., Choksi, M. Z., Barocas, S., De Sa, C., Grimmelmann, J., Kleinberg, J., Sen, S., & Zhang, B. (2024). Arbitrariness and Social Prediction: The Confounding Role of Variance in Fair Classification. Proceedings of the AAAI Conference on Artificial Intelligence, 38(20), 22004-22012. https://doi.org/10.1609/aaai.v38i20.30203