Covered Information Disentanglement: Model Transparency via Unbiased Permutation Importance

Authors

  • João P. B. Pereira Amsterdam UMC, Amsterdam, NL Horaizon B.V., Delft, NL
  • Erik S. G. Stroes Amsterdam UMC, Amsterdam, NL
  • Aeilko H. Zwinderman Amsterdam UMC, Amsterdam, NL
  • Evgeni Levin Amsterdam UMC, Amsterdam, NL Horaizon B.V., Delft, NL

DOI:

https://doi.org/10.1609/aaai.v36i7.20769

Keywords:

Machine Learning (ML)

Abstract

Model transparency is a prerequisite in many domains and an increasingly popular area in machine learning research. In the medical domain, for instance, unveiling the mechanisms behind a disease often has higher priority than the diagnostic itself since it might dictate or guide potential treatments and research directions. One of the most popular approaches to explain model global predictions is the permutation importance where the performance on permuted data is benchmarked against the baseline. However, this method and other related approaches will undervalue the importance of a feature in the presence of covariates since these cover part of its provided information. To address this issue, we propose Covered Information Disentanglement CID, a framework that considers all feature information overlap to correct the values provided by permutation importance. We further show how to compute CID efficiently when coupled with Markov random fields. We demonstrate its efficacy in adjusting permutation importance first on a controlled toy dataset and discuss its effect on real-world medical data.

Downloads

Published

2022-06-28

How to Cite

Pereira, J. P. B., Stroes, E. S. G., Zwinderman, A. H., & Levin, E. (2022). Covered Information Disentanglement: Model Transparency via Unbiased Permutation Importance. Proceedings of the AAAI Conference on Artificial Intelligence, 36(7), 7984-7992. https://doi.org/10.1609/aaai.v36i7.20769

Issue

Section

AAAI Technical Track on Machine Learning II