DiDA: Disambiguated Domain Alignment for Cross-Domain Retrieval with Partial Labels
DOI:
https://doi.org/10.1609/aaai.v38i4.28150Keywords:
CV: Image and Video Retrieval, ML: Multi-instance/Multi-view LearningAbstract
Driven by generative AI and the Internet, there is an increasing availability of a wide variety of images, leading to the significant and popular task of cross-domain image retrieval. To reduce annotation costs and increase performance, this paper focuses on an untouched but challenging problem, i.e., cross-domain image retrieval with partial labels (PCIR). Specifically, PCIR faces great challenges due to the ambiguous supervision signal and the domain gap. To address these challenges, we propose a novel method called disambiguated domain alignment (DiDA) for cross-domain retrieval with partial labels. In detail, DiDA elaborates a novel prototype-score unitization learning mechanism (PSUL) to extract common discriminative representations by simultaneously disambiguating the partial labels and narrowing the domain gap. Additionally, DiDA proposes a prototype-based domain alignment mechanism (PBDA) to further bridge the inherent cross-domain discrepancy. Attributed to PSUL and PBDA, our DiDA effectively excavates domain-invariant discrimination for cross-domain image retrieval. We demonstrate the effectiveness of DiDA through comprehensive experiments on three benchmarks, comparing it to existing state-of-the-art methods. Code available: https://github.com/lhrrrrrr/DiDA.Downloads
Published
2024-03-24
How to Cite
Liu, H., Ma, Y., Yan, M., Chen, Y., Peng, D., & Wang, X. (2024). DiDA: Disambiguated Domain Alignment for Cross-Domain Retrieval with Partial Labels. Proceedings of the AAAI Conference on Artificial Intelligence, 38(4), 3612-3620. https://doi.org/10.1609/aaai.v38i4.28150
Issue
Section
AAAI Technical Track on Computer Vision III