Reinforcement Based Learning on Classification Task Yields Better Generalization and Adversarial Accuracy (Student Abstract)

Authors

  • Shashi Kant Gupta Indian Institute of Technology Kanpur

DOI:

https://doi.org/10.1609/aaai.v35i18.17893

Keywords:

Adversarial Learning & Robustness, Robust Optimization, Adversarial Examples, Image Classification, Reinforcement Learning

Abstract

Deep Learning has become interestingly popular in the field of computer vision, mostly attaining near or above human-level performance in various vision tasks. But recent work has also demonstrated that these deep neural networks are very vulnerable to adversarial examples (adversarial examples - inputs to a model which are naturally similar to original data but fools the model in classifying it into a wrong class). In this work, we proposed a novel method to train deep learning models on an image classification task. We used a reward-based optimization function, similar to the vanilla policy gradient method in reinforcement learning to train our model instead of conventional cross-entropy loss. An empirical evaluation on cifar10 dataset showed that our method outperforms the same model architecture trained using cross-entropy loss function (on adversarial training). At the same time, our method generalizes better to the training data with the difference in test accuracy and train accuracy < 2% for most of the time as compared to cross-entropy one, whose difference most of the time remains > 2%.

Downloads

Published

2021-05-18

How to Cite

Gupta, S. K. (2021). Reinforcement Based Learning on Classification Task Yields Better Generalization and Adversarial Accuracy (Student Abstract). Proceedings of the AAAI Conference on Artificial Intelligence, 35(18), 15793-15794. https://doi.org/10.1609/aaai.v35i18.17893

Issue

Section

AAAI Student Abstract and Poster Program