DC-SPAN: A Dual Contrastive Attention Network for Multi-View Clustering
DOI:
https://doi.org/10.1609/aaai.v40i24.39099Abstract
Multi-view clustering aims to group data by integrating complementary information from multiple views. However, the inherent heterogeneity among views often leads to feature entanglement, severely limiting clustering performance. To address this challenge, we propose DC-SPAN—a Dual Contrastive Attention Network—grounded in a disentangle-then-fuse paradigm. DC-SPAN employs a dual-path variational architecture to explicitly decompose each view into shared and private latent subspaces. These representations are then robustly integrated via a Product-of-Experts (PoE) mechanism. At the heart of our model is a novel dual contrastive learning objective that simultaneously encourages alignment of shared components across views and enforces separation of private ones, enabling structured and disentangled representations. A gated attention fusion module further adaptively aggregates these latent factors to yield a unified, discriminative embedding. The overall model is trained end-to-end using a composite loss function that incorporates reconstruction, orthogonality, and contrastive terms, along with a two-stage training scheme for improved stability. Extensive experiments on benchmark datasets demonstrate that DC-SPAN consistently outperforms existing state-of-the-art methods, highlighting its effectiveness and robustness in handling multi-view heterogeneity.Downloads
Published
2026-03-14
How to Cite
Chen, J., Dong, Z., Li, T., & Han, Y. (2026). DC-SPAN: A Dual Contrastive Attention Network for Multi-View Clustering. Proceedings of the AAAI Conference on Artificial Intelligence, 40(24), 20127-20135. https://doi.org/10.1609/aaai.v40i24.39099
Issue
Section
AAAI Technical Track on Machine Learning I