Constraint-Driven Explanations for Black-Box ML Models

Authors

  • Aditya A. Shrotri Rice University
  • Nina Narodytska VMware Research Inc.
  • Alexey Ignatiev Monash University
  • Kuldeep S Meel National University of Singapore
  • Joao Marques-Silva IRIT, CNRS,Toulouse, France
  • Moshe Y. Vardi Rice University

DOI:

https://doi.org/10.1609/aaai.v36i8.20805

Keywords:

Machine Learning (ML), Constraint Satisfaction And Optimization (CSO), Reasoning Under Uncertainty (RU), Humans And AI (HAI)

Abstract

The need to understand the inner workings of opaque Machine Learning models has prompted researchers to devise various types of post-hoc explanations. A large class of such explainers proceed in two phases: first perturb an input instance whose explanation is sought, and then generate an interpretable artifact to explain the prediction of the opaque model on that instance. Recently, Deutch and Frost proposed to use an additional input from the user: a set of constraints over the input space to guide the perturbation phase. While this approach affords the user the ability to tailor the explanation to their needs, striking a balance between flexibility, theoretical rigor and computational cost has remained an open challenge. We propose a novel constraint-driven explanation generation approach which simultaneously addresses these issues in a modular fashion. Our framework supports the use of expressive Boolean constraints giving the user more flexibility to specify the subspace to generate perturbations from. Leveraging advances in Formal Methods, we can theoretically guarantee strict adherence of the samples to the desired distribution. This also allows us to compute fidelity in a rigorous way, while scaling much better in practice. Our empirical study demonstrates concrete uses of our tool CLIME in obtaining more meaningful explanations with high fidelity.

Downloads

Published

2022-06-28

How to Cite

Shrotri, A. A., Narodytska, N., Ignatiev, A., Meel, K. S., Marques-Silva, J., & Vardi, M. Y. (2022). Constraint-Driven Explanations for Black-Box ML Models. Proceedings of the AAAI Conference on Artificial Intelligence, 36(8), 8304-8314. https://doi.org/10.1609/aaai.v36i8.20805

Issue

Section

AAAI Technical Track on Machine Learning III