Dynamic Sub-graph Distillation for Robust Semi-supervised Continual Learning

Authors

  • Yan Fan Tianjin Key Lab of Machine Learning, College of Intelligence and Computing, Tianjin University Haihe Laboratory of Information Technology Application Innovation
  • Yu Wang Tianjin Key Lab of Machine Learning, College of Intelligence and Computing, Tianjin University Haihe Laboratory of Information Technology Application Innovation
  • Pengfei Zhu Tianjin Key Lab of Machine Learning, College of Intelligence and Computing, Tianjin University Haihe Laboratory of Information Technology Application Innovation
  • Qinghua Hu Tianjin Key Lab of Machine Learning, College of Intelligence and Computing, Tianjin University Haihe Laboratory of Information Technology Application Innovation

DOI:

https://doi.org/10.1609/aaai.v38i11.29079

Keywords:

ML: Life-Long and Continual Learning, ML: Semi-Supervised Learning

Abstract

Continual learning (CL) has shown promising results and comparable performance to learning at once in a fully supervised manner. However, CL strategies typically require a large number of labeled samples, making their real-life deployment challenging. In this work, we focus on semi-supervised continual learning (SSCL), where the model progressively learns from partially labeled data with unknown categories. We provide a comprehensive analysis of SSCL and demonstrate that unreliable distributions of unlabeled data lead to unstable training and refinement of the progressing stages. This problem severely impacts the performance of SSCL. To address the limitations, we propose a novel approach called Dynamic Sub-Graph Distillation (DSGD) for semi-supervised continual learning, which leverages both semantic and structural information to achieve more stable knowledge distillation on unlabeled data and exhibit robustness against distribution bias. Firstly, we formalize a general model of structural distillation and design a dynamic graph construction for the continual learning progress. Next, we define a structure distillation vector and design a dynamic sub-graph distillation algorithm, which enables end-to-end training and adaptability to scale up tasks. The entire proposed method is adaptable to various CL methods and supervision settings. Finally, experiments conducted on three datasets CIFAR10, CIFAR100, and ImageNet-100, with varying supervision ratios, demonstrate the effectiveness of our proposed approach in mitigating the catastrophic forgetting problem in semi-supervised continual learning scenarios. Our code is available: https://github.com/fanyan0411/DSGD.

Published

2024-03-24

How to Cite

Fan, Y., Wang, Y., Zhu, P., & Hu, Q. (2024). Dynamic Sub-graph Distillation for Robust Semi-supervised Continual Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 38(11), 11927-11935. https://doi.org/10.1609/aaai.v38i11.29079

Issue

Section

AAAI Technical Track on Machine Learning II