Multi-Level Confidence Learning for Trustworthy Multimodal Classification

Authors

  • Xiao Zheng National University of Defense Technology
  • Chang Tang China University of Geosciences
  • Zhiguo Wan Zhejiang Lab
  • Chengyu Hu China University of Geosciences
  • Wei Zhang Shandong Computer Science Center (National Supercomputing Center in Jinan)

DOI:

https://doi.org/10.1609/aaai.v37i9.26346

Keywords:

ML: Clustering, APP: Bioinformatics, ML: Classification and Regression, ML: Dimensionality Reduction/Feature Selection, ML: Multi-Instance/Multi-View Learning, ML: Multimodal Learning

Abstract

With the rapid development of various data acquisition technologies, more and more multimodal data come into being. It is important to integrate different modalities which are with high-dimensional features for boosting final multimodal data classification task. However, existing multimodal classification methods mainly focus on exploiting the complementary information of different modalities, while ignoring the learning confidence during information fusion. In this paper, we propose a trustworthy multimodal classification network via multi-level confidence learning, referred to as MLCLNet. Considering that a large number of feature dimensions could not contribute to final classification performance but disturb the discriminability of different samples, we propose a feature confidence learning mechanism to suppress some redundant features, as well as enhancing the expression of discriminative feature dimensions in each modality. In order to capture the inherent sample structure information implied in each modality, we design a graph convolutional network branch to learn the corresponding structure preserved feature representation and generate modal-specific initial classification labels. Since samples from different modalities should share consistent labels, a cross-modal label fusion module is deployed to capture the label correlations of different modalities. In addition, motivated the ideally orthogonality of final fused label matrix, we design a label confidence loss to supervise the network for learning more separable data representations. To the best of our knowledge, MLCLNet is the first work which integrates both feature and label-level confidence learning for multimodal classification. Extensive experiments on four multimodal medical datasets are conducted to validate superior performance of MLCLNet when compared to other state-of-the-art methods.

Downloads

Published

2023-06-26

How to Cite

Zheng, X., Tang, C., Wan, Z., Hu, C., & Zhang, W. (2023). Multi-Level Confidence Learning for Trustworthy Multimodal Classification. Proceedings of the AAAI Conference on Artificial Intelligence, 37(9), 11381-11389. https://doi.org/10.1609/aaai.v37i9.26346

Issue

Section

AAAI Technical Track on Machine Learning IV