Delivering Inflated Explanations

Authors

  • Yacine Izza CREATE, National University of Singapore
  • Alexey Ignatiev Monash University
  • Peter J. Stuckey Monash University OPTIMA ARC Industrial Training and Transformation Centre
  • Joao Marques-Silva IRIT, CNRS

DOI:

https://doi.org/10.1609/aaai.v38i11.29170

Keywords:

ML: Transparent, Interpretable, Explainable ML, CSO: Satisfiability, CSO: Constraint Satisfaction, KRR: Diagnosis and Abductive Reasoning, KRR: Automated Reasoning and Theorem Proving

Abstract

In the quest for Explainable Artificial Intelligence (XAI) one of the questions that frequently arises given a decision made by an AI system is, ``why was the decision made in this way?'' Formal approaches to explainability build a formal model of the AI system and use this to reason about the properties of the system. Given a set of feature values for an instance to be explained, and a resulting decision, a formal abductive explanation is a set of features, such that if they take the given value will always lead to the same decision. This explanation is useful, it shows that only some features were used in making the final decision. But it is narrow, it only shows that if the selected features take their given values the decision is unchanged. It is possible that some features may change values and still lead to the same decision. In this paper we formally define inflated explanations which is a set of features, and for each feature a set of values (always including the value of the instance being explained), such that the decision will remain unchanged, for any of the values allowed for any of the features in the (inflated) abductive explanation. Inflated formal explanations are more informative than common abductive explanations since e.g. they allow us to see if the exact value of a feature is important, or it could be any nearby value. Overall they allow us to better understand the role of each feature in the decision. We show that we can compute inflated explanations for not that much greater cost than abductive explanations, and that we can extend duality results for abductive explanations also to inflated explanations.

Published

2024-03-24

How to Cite

Izza, Y., Ignatiev, A., Stuckey, P. J., & Marques-Silva, J. (2024). Delivering Inflated Explanations. Proceedings of the AAAI Conference on Artificial Intelligence, 38(11), 12744-12753. https://doi.org/10.1609/aaai.v38i11.29170

Issue

Section

AAAI Technical Track on Machine Learning II