Visual Chain-of-Thought Prompting for Knowledge-Based Visual Reasoning

Authors

  • Zhenfang Chen MIT-IBM Watson AI Lab
  • Qinhong Zhou UMass Amherst
  • Yikang Shen MIT-IBM Watson AI Lab
  • Yining Hong University of California, Los Angeles
  • Zhiqing Sun Carnegie Mellon University
  • Dan Gutfreund MIT-IBM Watson AI Lab
  • Chuang Gan MIT-IBM Watson AI Lab UMass Amherst

DOI:

https://doi.org/10.1609/aaai.v38i2.27888

Keywords:

CV: Language and Vision, CV: Visual Reasoning & Symbolic Representations

Abstract

Knowledge-based visual reasoning remains a daunting task since it not only requires machines to interpret the concepts and relationships from visual scenes but also associate them with external world knowledge to conduct a chain of reasoning on open-world questions. Previous works, however, treat visual perception and language-based reasoning as two independent modules, failing to attend to both modules throughout all stages of reasoning. To this end, we propose Visual Chain-of-thought Prompting (VCTP) for knowledge-based reasoning, which involves the interaction between visual content and natural language in an iterative step-by-step reasoning manner. VCTP contains three stages, see, think, and confirm. The see stage scans the image and grounds the visual concept candidates with a visual perception model. The think stage adopts a pre-trained large language model (LLM) to attend to key visual concepts from natural language questions adaptively. It then transforms key visual context into text context for prompting with a visual captioning model, and adopts the LLM to generate the answer. The confirm stage further uses the LLM to generate the supporting rationale to the answer, which is then passed through a cross-modality classifier to verify that it’s consistent with the visual context. We iterate through the think-confirm stages to ensure the verified rationale is consistent with the answer. We conduct experiments on a range of knowledge-based visual reasoning datasets. We found our VCTP enjoys several benefits, 1). it achieves better performance than the previous few-shot learning baselines; 2). it enjoys the total transparency and trustworthiness of the whole reasoning process by providing rationales for each reasoning step; 3). it is computation-efficient compared with other fine-tuning baselines. Our code is available at https://github.com/UMass-Foundation-Model/VisualCoT.git

Published

2024-03-24

How to Cite

Chen, Z., Zhou, Q., Shen, Y., Hong, Y., Sun, Z., Gutfreund, D., & Gan, C. (2024). Visual Chain-of-Thought Prompting for Knowledge-Based Visual Reasoning. Proceedings of the AAAI Conference on Artificial Intelligence, 38(2), 1254-1262. https://doi.org/10.1609/aaai.v38i2.27888

Issue

Section

AAAI Technical Track on Computer Vision I