Visual Explanations for Convolutional Neural Networks via Latent Traversal of Generative Adversarial Networks (Student Abstract)

Authors

  • Amil Dravid Northwestern University
  • Aggelos K. Katsaggelos Northwestern University

DOI:

https://doi.org/10.1609/aaai.v36i11.21606

Keywords:

Deep Learning, Generative Adversarial Networks, Explainable AI, Medical Imaging

Abstract

Lack of explainability in artificial intelligence, specifically deep neural networks, remains a bottleneck for implementing models in practice. Popular techniques such as Gradient-weighted Class Activation Mapping (Grad-CAM) provide a coarse map of salient features in an image, which rarely tells the whole story of what a convolutional neural network(CNN) learned. Using COVID-19 chest X-rays, we present a method for interpreting what a CNN has learned by utilizing Generative Adversarial Networks (GANs). Our GAN framework disentangles lung structure from COVID-19 features. Using this GAN, we can visualize the transition of a pair of COVID negative lungs in a chest radiograph to a COVID positive pair by interpolating in the latent space of the GAN, which provides fine-grained visualization of how the CNN responds to varying features within the lungs.

Downloads

Published

2022-06-28

How to Cite

Dravid, A., & Katsaggelos, A. K. (2022). Visual Explanations for Convolutional Neural Networks via Latent Traversal of Generative Adversarial Networks (Student Abstract). Proceedings of the AAAI Conference on Artificial Intelligence, 36(11), 12939-12940. https://doi.org/10.1609/aaai.v36i11.21606