Multi-Teacher Knowledge Distillation with Reinforcement Learning for Visual Recognition

Authors

  • Chuanguang Yang Institute of Computing Technology, Chinese Academy of Sciences
  • XinQiang Yu Institute of Computing Technology, Chinese Academy of Sciences University of Chinese Academy of Sciences
  • Han Yang Institute of Computing Technology, Chinese Academy of SciencesUniversity of Chinese Academy of Sciences
  • Zhulin An Institute of Computing Technology, Chinese Academy of Sciences
  • Chengqing Yu Institute of Computing Technology, Chinese Academy of Sciences University of Chinese Academy of Sciences
  • Libo Huang Institute of Computing Technology, Chinese Academy of Sciences
  • Yongjun Xu Institute of Computing Technology, Chinese Academy of Sciences

DOI:

https://doi.org/10.1609/aaai.v39i9.32990

Abstract

Multi-teacher Knowledge Distillation (KD) transfers diverse knowledge from a teacher pool to a student network. The core problem of multi-teacher KD is how to balance distillation strengths among various teachers. Most existing methods often develop weighting strategies from an individual perspective of teacher performance or teacher-student gaps, lacking comprehensive information for guidance. This paper proposes Multi-Teacher Knowledge Distillation with Reinforcement Learning (MTKD-RL) to optimize multi-teacher weights. In this framework, we construct both teacher performance and teacher-student gaps as state information to an agent. The agent outputs the teacher weight and can be updated by the return reward from the student. MTKD-RL reinforces the interaction between the student and teacher using an agent in an RL-based decision mechanism, achieving better matching capability with more meaningful weights. Experimental results on visual recognition tasks, including image classification, object detection, and semantic segmentation tasks, demonstrate that MTKD-RL achieves state-of-the-art performance compared to the existing multi-teacher KD works.

Published

2025-04-11

How to Cite

Yang, C., Yu, X., Yang, H., An, Z., Yu, C., Huang, L., & Xu, Y. (2025). Multi-Teacher Knowledge Distillation with Reinforcement Learning for Visual Recognition. Proceedings of the AAAI Conference on Artificial Intelligence, 39(9), 9148–9156. https://doi.org/10.1609/aaai.v39i9.32990

Issue

Section

AAAI Technical Track on Computer Vision VIII