Efficient Folded Attention for Medical Image Reconstruction and Segmentation

Authors

  • Hang Zhang Cornell University Weill Cornell Medical College
  • Jinwei Zhang Cornell University Weill Cornell Medical College
  • Rongguang Wang University of Pennsylvania
  • Qihao Zhang Cornell University Weill Cornell Medical College
  • Pascal Spincemaille Cornell University
  • Thanh D. Nguyen Cornell University
  • Yi Wang Cornell University Weill Cornell Medical College

DOI:

https://doi.org/10.1609/aaai.v35i12.17298

Keywords:

(Deep) Neural Network Algorithms, Applications, Healthcare, Medicine & Wellness, Segmentation

Abstract

Recently, 3D medical image reconstruction (MIR) and segmentation (MIS) based on deep neural networks have been developed with promising results, and attention mechanism has been further designed for performance enhancement. However, the large size of 3D volume images poses a great computational challenge to traditional attention methods. In this paper, we propose a folded attention (FA) approach to improve the computational efficiency of traditional attention methods on 3D medical images. The main idea is that we apply tensor folding and unfolding operations to construct four small sub-affinity matrices to approximate the original affinity matrix. Through four consecutive sub-attention modules of FA, each element in the feature tensor can aggregate spatial-channel information from all other elements. Compared to traditional attention methods, with the moderate improvement of accuracy, FA can substantially reduce the computational complexity and GPU memory consumption. We demonstrate the superiority of our method on two challenging tasks for 3D MIR and MIS, which are quantitative susceptibility mapping and multiple sclerosis lesion segmentation.

Downloads

Published

2021-05-18

How to Cite

Zhang, H., Zhang, J., Wang, R., Zhang, Q., Spincemaille, P., Nguyen, T. D., & Wang, Y. (2021). Efficient Folded Attention for Medical Image Reconstruction and Segmentation. Proceedings of the AAAI Conference on Artificial Intelligence, 35(12), 10868-10876. https://doi.org/10.1609/aaai.v35i12.17298

Issue

Section

AAAI Technical Track on Machine Learning V