Symbolic Metamodels for Interpreting Black-Boxes Using Primitive Functions

Authors

  • Mahed Abroshan The Alan Turing Institute, London, UK
  • Saumitra Mishra JP Morgan AI Research, London, UK
  • Mohammad Mahdi Khalili Yahoo! Research, NYC, NY, USA CSE Department, The Ohio State University, Columbus, Ohio, USA

DOI:

https://doi.org/10.1609/aaai.v37i6.25816

Keywords:

ML: Transparent, Interpretable, Explainable ML, PEAI: Interpretability and Explainability, SO: Evolutionary Computation

Abstract

One approach for interpreting black-box machine learning models is to find a global approximation of the model using simple interpretable functions, which is called a metamodel (a model of the model). Approximating the black-box with a metamodel can be used to 1) estimate instance-wise feature importance; 2) understand the functional form of the model; 3) analyze feature interactions. In this work, we propose a new method for finding interpretable metamodels. Our approach utilizes Kolmogorov superposition theorem, which expresses multivariate functions as a composition of univariate functions (our primitive parameterized functions). This composition can be represented in the form of a tree. Inspired by symbolic regression, we use a modified form of genetic programming to search over different tree configurations. Gradient descent (GD) is used to optimize the parameters of a given configuration. Our method is a novel memetic algorithm that uses GD not only for training numerical constants but also for the training of building blocks. Using several experiments, we show that our method outperforms recent metamodeling approaches suggested for interpreting black-boxes.

Downloads

Published

2023-06-26

How to Cite

Abroshan, M., Mishra, S., & Khalili, M. M. (2023). Symbolic Metamodels for Interpreting Black-Boxes Using Primitive Functions. Proceedings of the AAAI Conference on Artificial Intelligence, 37(6), 6649-6657. https://doi.org/10.1609/aaai.v37i6.25816

Issue

Section

AAAI Technical Track on Machine Learning I