Multi-Knowledge Aggregation and Transfer for Semantic Segmentation

Authors

  • Yuang Liu East China Normal University
  • Wei Zhang East China Normal University
  • Jun Wang East China Normal University

DOI:

https://doi.org/10.1609/aaai.v36i2.20077

Keywords:

Computer Vision (CV)

Abstract

As a popular deep neural networks (DNN) compression technique, knowledge distillation (KD) has attracted increasing attentions recently. Existing KD methods usually utilize one kind of knowledge in an intermediate layer of DNN for classification tasks to transfer useful information from cumbersome teacher networks to compact student networks. However, this paradigm is not very suitable for semantic segmentation, a comprehensive vision task based on both pixel-level and contextual information, since it cannot provide rich information for distillation. In this paper, we propose a novel multi-knowledge aggregation and transfer (MKAT) framework to comprehensively distill knowledge within an intermediate layer for semantic segmentation. Specifically, the proposed framework consists of three parts: Independent Transformers and Encoders module (ITE), Auxiliary Prediction Branch (APB), and Mutual Label Calibration (MLC) mechanism, which can take advantage of abundant knowledge from intermediate features. To demonstrate the effectiveness of our proposed approach, we conduct extensive experiments on three segmentation datasets: Pascal VOC, Cityscapes, and CamVid, showing that MKAT outperforms the other KD methods.

Downloads

Published

2022-06-28

How to Cite

Liu, Y., Zhang, W., & Wang, J. (2022). Multi-Knowledge Aggregation and Transfer for Semantic Segmentation. Proceedings of the AAAI Conference on Artificial Intelligence, 36(2), 1837-1845. https://doi.org/10.1609/aaai.v36i2.20077

Issue

Section

AAAI Technical Track on Computer Vision II