An Adversarial Framework for Generating Unseen Images by Activation Maximization

Authors

  • Yang Zhang MIT-IBM Watson AI Lab
  • Wang Zhou Meta AI
  • Gaoyuan Zhang MIT-IBM Watson AI Lab
  • David Cox MIT-IBM Watson AI Lab
  • Shiyu Chang University of California, Santa Barbara

DOI:

https://doi.org/10.1609/aaai.v36i3.20247

Keywords:

Computer Vision (CV)

Abstract

Activation maximization (AM) refers to the task of generating input examples that maximize the activation of a target class of a classifier, which can be used for class-conditional image generation and model interpretation. A popular class of AM method, GAN-based AM, introduces a GAN pre-trained on a large image set, and performs AM over its input random seed or style embeddings, so that the generated images are natural and adversarial attacks are prevented. Most of these methods would require the image set to contain some images of the target class to be visualized. Otherwise they tend to generate other seen class images that most maximizes the target class activation. In this paper, we aim to tackle the case where information about the target class is completely removed from the image set. This would ensure that the generated images truly reflect the target class information residing in the classifier, not the target class information in the image set, which contributes to a more faithful interpretation technique. To this end, we propose PROBEGAN, a GAN-based AM algorithm capable of generating image classes unseen in the image set. Rather than using a pre-trained GAN, PROBEGAN trains a new GAN with AM explicitly included in its training objective. PROBEGAN consists of a class-conditional generator, a seen-class discriminator, and an all-class unconditional discriminator. It can be shown that such a framework can generate images with the features of the unseen target class, while retaining the naturalness as depicted in the image set. Experiments have shown that PROBEGAN can generate unseen-class images with much higher quality than the baselines. We also explore using PROBEGAN as a model interpretation tool. Our code is at https://github.com/csmiler/ProbeGAN/.

Downloads

Published

2022-06-28

How to Cite

Zhang, Y., Zhou, W., Zhang, G., Cox, D., & Chang, S. (2022). An Adversarial Framework for Generating Unseen Images by Activation Maximization. Proceedings of the AAAI Conference on Artificial Intelligence, 36(3), 3371-3379. https://doi.org/10.1609/aaai.v36i3.20247

Issue

Section

AAAI Technical Track on Computer Vision III