Signal: Selective Interaction and Global-local Alignment for Multi-Modal Object Re-Identification
DOI:
https://doi.org/10.1609/aaai.v40i9.37674Abstract
Multi-modal object Re-IDentification (ReID) is devoted to retrieving specific objects through the exploitation of complementary multi-modal image information. Existing methods mainly concentrate on the fusion of multi-modal features, yet neglecting the background interference. Besides, current multi-modal fusion methods often focus on aligning modality pairs but suffer from multi-modal consistency alignment. To address these issues, we propose a novel selective interaction and global-local alignment framework called Signal for multi-modal object ReID. Specifically, we first propose a Selective Interaction Module (SIM) to select important patch tokens with intra-modal and inter-modal information. These important patch tokens engage in the interaction with class tokens, thereby yielding more discriminative features. Then, we propose a Global Alignment Module (GAM) to simultaneously align multi-modal features by minimizing the volume of 3D polyhedra in the gramian space. Meanwhile, we propose a Local Alignment Module (LAM) to align local features in a shift-aware manner. With these modules, our proposed framework could extract more discriminative features for object ReID. Extensive experiments on three multi-modal object ReID benchmarks (i.e., RGBNT201, RGBNT100, MSVR310) validate the effectiveness of our method.Published
2026-03-14
How to Cite
Liu, Y., Wang, Y., & Zhang, P. (2026). Signal: Selective Interaction and Global-local Alignment for Multi-Modal Object Re-Identification. Proceedings of the AAAI Conference on Artificial Intelligence, 40(9), 7359–7367. https://doi.org/10.1609/aaai.v40i9.37674
Issue
Section
AAAI Technical Track on Computer Vision VI