Very Fast, Approximate Counterfactual Explanations for Decision Forests

Authors

  • Miguel Á. Carreira-Perpinan UC Merced
  • Suryabhan Singh Hada UC Merced

DOI:

https://doi.org/10.1609/aaai.v37i6.25848

Keywords:

ML: Transparent, Interpretable, Explainable ML, ML: Bias and Fairness, ML: Ensemble Methods, PEAI: Interpretability and Explainability

Abstract

We consider finding a counterfactual explanation for a classification or regression forest, such as a random forest. This requires solving an optimization problem to find the closest input instance to a given instance for which the forest outputs a desired value. Finding an exact solution has a cost that is exponential on the number of leaves in the forest. We propose a simple but very effective approach: we constrain the optimization to input space regions populated by actual data points. The problem reduces to a form of nearest-neighbor search using a certain distance on a certain dataset. This has two advantages: first, the solution can be found very quickly, scaling to large forests and high-dimensional data, and enabling interactive use. Second, the solution found is more likely to be realistic in that it is guided towards high-density areas of input space.

Downloads

Published

2023-06-26

How to Cite

Carreira-Perpinan, M. Á., & Hada, S. S. (2023). Very Fast, Approximate Counterfactual Explanations for Decision Forests. Proceedings of the AAAI Conference on Artificial Intelligence, 37(6), 6935-6943. https://doi.org/10.1609/aaai.v37i6.25848

Issue

Section

AAAI Technical Track on Machine Learning I