Manipulation Intention Understanding for Zero-Shot Composed Image Retrieval
DOI:
https://doi.org/10.1609/aaai.v40i11.37907Abstract
Zero-shot Composed Image Retrieval (ZS-CIR) involves diverse tasks with varied visual manipulation intents across domains, scenes, objects, and attributes. A key challenge is that existing datasets contain limited intent-relevant annotations, making it hard for models to infer human intent from textual modifications. We introduce an intent-centric image–text dataset generated via reasoning by a Multimodal Large Language Model (MLLM) to better train ZS-CIR models for human manipulation intent understanding. Building on this dataset, we propose De-MINDS, a framework that distills the MLLM’s reasoning ability to capture manipulation intent and enhance models’ comprehension of modified text. A simple mapping network translates image information into language space and combines it with the manipulation text to form a query. De-MINDS then extracts intention-relevant information from this query and encodes it as pseudo-word tokens for accurate ZS-CIR. Across four ZS-CIR tasks, De-MINDS shows strong generalization and improves over existing methods by 2.15% to 4.05%, establishing new state-of-the-art results with comparable inference time.Published
2026-03-14
How to Cite
Tang, Y., Yu, J., Gai, K., Xiong, G., Gou, G., Qiu, M., & Wu, Q. (2026). Manipulation Intention Understanding for Zero-Shot Composed Image Retrieval. Proceedings of the AAAI Conference on Artificial Intelligence, 40(11), 9466-9474. https://doi.org/10.1609/aaai.v40i11.37907
Issue
Section
AAAI Technical Track on Computer Vision VIII