Enhancing Multimodal Large Language Models Complex Reason via Similarity Computation

Authors

  • Xiaofeng Zhang Shanghai Jiaotong University
  • Fanshuo Zeng Institute of Automation,Chinese Academy of Sciences
  • Yihao Quan Beijing Jiaotong University
  • Zheng Hui Columbia University
  • Jiawei Yao University of Washington

DOI:

https://doi.org/10.1609/aaai.v39i10.33107

Abstract

Multimodal large language models have experienced rapid growth, and numerous different models have emerged. The interpretability of LVLMs remains an under-explored area. Especially when faced with more complex tasks such as chain-of-thought reasoning, its internal mechanisms still resemble a black box that is difficult to decipher. By studying the interaction and information flow between images and text, we noticed that in models such as LLaVA1.5, image tokens that are semantically related to text are more likely to have information flow convergence in the LLM decoding layer, and these image tokens receive higher attention scores. However, those image tokens that are less relevant to the text do not have information flow convergence, and they only get very small attention scores. To efficiently utilize the image information, we propose a new image token reduction method, Simignore, which aims to improve the complex reasoning ability of LVLMs by computing the similarity between image and text embeddings and ignoring image tokens that are irrelevant and unimportant to the text. Through extensive experiments, we demonstrate the effectiveness of our method for complex reasoning tasks.

Downloads

Published

2025-04-11

How to Cite

Zhang, X., Zeng, F., Quan, Y., Hui, Z., & Yao, J. (2025). Enhancing Multimodal Large Language Models Complex Reason via Similarity Computation. Proceedings of the AAAI Conference on Artificial Intelligence, 39(10), 10203-10211. https://doi.org/10.1609/aaai.v39i10.33107

Issue

Section

AAAI Technical Track on Computer Vision IX