Text-Guided Attention Model for Image Captioning

Authors

  • Jonghwan Mun Pohang University of Science and Technology (POSTECH)
  • Minsu Cho Pohang University of Science and Technology (POSTECH)
  • Bohyung Han Pohang University of Science and Technology (POSTECH)

DOI:

https://doi.org/10.1609/aaai.v31i1.11237

Keywords:

Image Captioning, Attention Model

Abstract

Visual attention plays an important role to understand images and demonstrates its effectiveness in generating natural language descriptions of images. On the other hand, recent studies show that language associated with an image can steer visual attention in the scene during our cognitive process. Inspired by this, we introduce a text-guided attention model for image captioning, which learns to drive visual attention using associated captions. For this model, we propose an exemplar-based learning approach that retrieves from training data associated captions with each image, and use them to learn attention on visual features. Our attention model enables to describe a detailed state of scenes by distinguishing small or confusable objects effectively. We validate our model on MS-COCO Captioning benchmark and achieve the state-of-the-art performance in standard metrics.

Downloads

Published

2017-02-12

How to Cite

Mun, J., Cho, M., & Han, B. (2017). Text-Guided Attention Model for Image Captioning. Proceedings of the AAAI Conference on Artificial Intelligence, 31(1). https://doi.org/10.1609/aaai.v31i1.11237