Order-Free RNN With Visual Attention for Multi-Label Classification

Authors

  • Shang-Fu Chen National Taiwan University
  • Yi-Chen Chen National Taiwan University
  • Chih-Kuan Yeh Carnegie Mellon University
  • Yu-Chiang Wang National Taiwan University

DOI:

https://doi.org/10.1609/aaai.v32i1.12230

Keywords:

multi-label

Abstract

We propose a recurrent neural network (RNN) based model for image multi-label classification. Our model uniquely integrates and learning of visual attention and Long Short Term Memory (LSTM) layers, which jointly learns the labels of interest and their co-occurrences, while the associated image regions are visually attended. Different from existing approaches utilize either model in their network architectures, training of our model does not require pre-defined label orders. Moreover, a robust inference process is introduced so that prediction errors would not propagate and thus affect the performance. Our experiments on NUS-WISE and MS-COCO datasets confirm the design of our network and its effectiveness in solving multi-label classification problems.

Downloads

Published

2018-04-27

How to Cite

Chen, S.-F., Chen, Y.-C., Yeh, C.-K., & Wang, Y.-C. (2018). Order-Free RNN With Visual Attention for Multi-Label Classification. Proceedings of the AAAI Conference on Artificial Intelligence, 32(1). https://doi.org/10.1609/aaai.v32i1.12230