Searching for Unfairness in Algorithms’ Outputs: Novel Tests and Insights

Authors

  • Ian Davidson University of California, Davis
  • S. S. Ravi University of Virginia

DOI:

https://doi.org/10.1609/aaai.v39i26.34933

Abstract

As AI algorithms are deployed extensively, the need to ensure the fairness of their outputs is critical. Most existing work is on “fairness by design” approaches that incorporate limited tests for fairness into a limited number of algorithms. Here, we explore a framework that removes these limitations and can be used with any algorithm’s output that allocates instances to one of K categories/classes such as outlier detection (OD), clustering and classification. The framework can encode standard and novel fairness types beyond simple counting, and importantly, it can detect intersectional unfairness without being specifically told what to look for. Our experimental results show that both standard and novel types of unfairness exist extensively in the outputs of fair-by-design algorithms and the counter-intuitive result that they can actually increase intersectional unfairness.

Downloads

Published

2025-04-11

How to Cite

Davidson, I., & Ravi, S. S. (2025). Searching for Unfairness in Algorithms’ Outputs: Novel Tests and Insights. Proceedings of the AAAI Conference on Artificial Intelligence, 39(26), 27242–27249. https://doi.org/10.1609/aaai.v39i26.34933

Issue

Section

AAAI Technical Track on AI Alignment