ACM: Adaptive Cross-Modal Graph Convolutional Neural Networks for RGB-D Scene Recognition


  • Yuan Yuan Northwestern Polytechnical University
  • Zhitong Xiong Northwestern Polytechnical University
  • Qi Wang Northwestern Polytechnical University



RGB image classification has achieved significant performance improvement with the resurge of deep convolutional neural networks. However, mono-modal deep models for RGB image still have several limitations when applied to RGB-D scene recognition. 1) Images for scene classification usually contain more than one typical object with flexible spatial distribution, so the object-level local features should also be considered in addition to global scene representation. 2) Multi-modal features in RGB-D scene classification are still under-utilized. Simply combining these modal-specific features suffers from the semantic gaps between different modalities. 3) Most existing methods neglect the complex relationships among multiple modality features. Considering these limitations, this paper proposes an adaptive crossmodal (ACM) feature learning framework based on graph convolutional neural networks for RGB-D scene recognition. In order to make better use of the modal-specific cues, this approach mines the intra-modality relationships among the selected local features from one modality. To leverage the multi-modal knowledge more effectively, the proposed approach models the inter-modality relationships between two modalities through the cross-modal graph (CMG). We evaluate the proposed method on two public RGB-D scene classification datasets: SUN-RGBD and NYUD V2, and the proposed method achieves state-of-the-art performance.




How to Cite

Yuan, Y., Xiong, Z., & Wang, Q. (2019). ACM: Adaptive Cross-Modal Graph Convolutional Neural Networks for RGB-D Scene Recognition. Proceedings of the AAAI Conference on Artificial Intelligence, 33(01), 9176-9184.



AAAI Technical Track: Vision