“If it didn’t happen, why would I change my decision?”: How Judges Respond to Counterfactual Explanations for the Public Safety Assessment

Authors

  • Yaniv Yacoby Harvard University
  • Ben Green University of Michigan
  • Christopher L. Griffin Jr. James E. Rogers College of Law, University of Arizona
  • Finale Doshi-Velez Harvard University

DOI:

https://doi.org/10.1609/hcomp.v10i1.22001

Keywords:

Explainable AI, Counterfactual Explanations, Pretrial Risk Assessment, Criminal Justice

Abstract

Many researchers and policymakers have expressed excitement about algorithmic explanations enabling more fair and responsible decision-making. However, recent experimental studies have found that explanations do not always improve human use of algorithmic advice. In this study, we shed light on how people interpret and respond to counterfactual explanations (CFEs)---explanations that show how a model's output would change with marginal changes to its input(s)---in the context of pretrial risk assessment instruments (PRAIs). We ran think-aloud trials with eight sitting U.S. state court judges, providing them with recommendations from a PRAI that includes CFEs. We found that the CFEs did not alter the judges' decisions. At first, judges misinterpreted the counterfactuals as real---rather than hypothetical---changes to defendants. Once judges understood what the counterfactuals meant, they ignored them, stating their role is only to make decisions regarding the actual defendant in question. The judges also expressed a mix of reasons for ignoring or following the advice of the PRAI without CFEs. These results add to the literature detailing the unexpected ways in which people respond to algorithms and explanations. They also highlight new challenges associated with improving human-algorithm collaborations through explanations.

Downloads

Published

2022-10-14

How to Cite

Yacoby, Y., Green, B., Griffin Jr., C. L., & Doshi-Velez, F. (2022). “If it didn’t happen, why would I change my decision?”: How Judges Respond to Counterfactual Explanations for the Public Safety Assessment. Proceedings of the AAAI Conference on Human Computation and Crowdsourcing, 10(1), 219-230. https://doi.org/10.1609/hcomp.v10i1.22001