CLUENet: Cluster Attention Makes Neural Networks Have Eyes
DOI:
https://doi.org/10.1609/aaai.v40i11.37867Abstract
Despite the success of convolution- and attention-based models in vision tasks, their rigid receptive fields and complex architectures limit their ability to model irregular spatial patterns and hinder interpretability, thereby posing challenges for tasks requiring high model transparency. Clustering paradigms offer promising interpretability and flexible semantic modeling, but suffer from limited accuracy, low efficiency, and gradient vanishing during training. To address these issues, we propose the CLUster attEntion Network (CLUENet), a transparent deep architecture for visual semantic understanding. Specifically, we introduce three key innovations, including (i) a Global and Soft Feature Aggregation with a Temperature-Scaled Cosine Attention for capturing long-range dependencies and a Gated Fusion Mechanism for enhanced local modeling, (ii) Hard and Shared Feature Dispatching, and (iii) an Improved Cluster Pooling Block. These enhancements significantly improve both classification performance and visual interpretability. Experiments on CIFAR-100 and Mini-ImageNet demonstrate that CLUENet outperforms existing clustering methods and mainstream visual models, offering a compelling balance of accuracy, efficiency, and transparency.Downloads
Published
2026-03-14
How to Cite
Song, X., Huang, J.-J., Liu, T., Liang, K., & Tang, C. (2026). CLUENet: Cluster Attention Makes Neural Networks Have Eyes. Proceedings of the AAAI Conference on Artificial Intelligence, 40(11), 9106–9115. https://doi.org/10.1609/aaai.v40i11.37867
Issue
Section
AAAI Technical Track on Computer Vision VIII