Towards Continual Learning Desiderata via HSIC-Bottleneck Orthogonalization and Equiangular Embedding

Authors

  • Depeng Li School of Artificial Intelligence and Automation, Huazhong University of Science and Technology
  • Tianqi Wang School of Artificial Intelligence and Automation, Huazhong University of Science and Technology
  • Junwei Chen School of Artificial Intelligence and Automation, Huazhong University of Science and Technology
  • Qining Ren School of Artificial Intelligence and Automation, Huazhong University of Science and Technology
  • Kenji Kawaguchi School of Computing, National University of Singapore
  • Zhigang Zeng School of Artificial Intelligence and Automation, Huazhong University of Science and Technology

DOI:

https://doi.org/10.1609/aaai.v38i12.29249

Keywords:

ML: Life-Long and Continual Learning, ML: Time-Series/Data Streams, ML: Classification and Regression, ML: Other Foundations of Machine Learning

Abstract

Deep neural networks are susceptible to catastrophic forgetting when trained on sequential tasks. Various continual learning (CL) methods often rely on exemplar buffers or/and network expansion for balancing model stability and plasticity, which, however, compromises their practical value due to privacy and memory concerns. Instead, this paper considers a strict yet realistic setting, where the training data from previous tasks is unavailable and the model size remains relatively constant during sequential training. To achieve such desiderata, we propose a conceptually simple yet effective method that attributes forgetting to layer-wise parameter overwriting and the resulting decision boundary distortion. This is achieved by the synergy between two key components: HSIC-Bottleneck Orthogonalization (HBO) implements non-overwritten parameter updates mediated by Hilbert-Schmidt independence criterion in an orthogonal space and EquiAngular Embedding (EAE) enhances decision boundary adaptation between old and new tasks with predefined basis vectors. Extensive experiments demonstrate that our method achieves competitive accuracy performance, even with absolute superiority of zero exemplar buffer and 1.02x the base model.

Published

2024-03-24

How to Cite

Li, D., Wang, T., Chen, J., Ren, Q., Kawaguchi, K., & Zeng, Z. (2024). Towards Continual Learning Desiderata via HSIC-Bottleneck Orthogonalization and Equiangular Embedding. Proceedings of the AAAI Conference on Artificial Intelligence, 38(12), 13464-13473. https://doi.org/10.1609/aaai.v38i12.29249

Issue

Section

AAAI Technical Track on Machine Learning III