Pareto Continual Learning: Preference-Conditioned Learning and Adaption for Dynamic Stability-Plasticity Trade-off

Authors

  • Song Lai Department of Computer Science, City University of Hong Kong Centre for Artificial Intelligence and Robotics, HK Institute of Science & Innovation, Chinese Academy of Sciences City University of Hong Kong Shenzhen Research Institute
  • Zhe Zhao Department of Computer Science, City University of Hong Kong University of Science and Technology of China
  • Fei Zhu Centre for Artificial Intelligence and Robotics, HK Institute of Science & Innovation, Chinese Academy of Sciences
  • Xi Lin Department of Computer Science, City University of Hong Kong
  • Qingfu Zhang Department of Computer Science, City University of Hong Kong City University of Hong Kong Shenzhen Research Institute
  • Gaofeng Meng Centre for Artificial Intelligence and Robotics, HK Institute of Science & Innovation, Chinese Academy of Sciences State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences

DOI:

https://doi.org/10.1609/aaai.v39i17.33981

Abstract

Continual learning aims to learn multiple tasks sequentially. A key challenge in continual learning is balancing between two objectives: retaining knowledge from old tasks (stability) and adapting to new tasks (plasticity). Experience replay methods, which store and replay past data alongside new data, have become a widely adopted approach to mitigate catastrophic forgetting. However, these methods neglect the dynamic nature of the stability-plasticity trade-off and aim to find a fixed and unchanging balance, resulting in suboptimal adaptation during training and inference. In this paper, we propose Pareto Continual Learning (ParetoCL), a novel framework that reformulates the stability-plasticity trade-off in continual learning as a multi-objective optimization (MOO) problem. ParetoCL introduces a preference-conditioned model to efficiently learn a set of Pareto optimal solutions representing different trade-offs and enables dynamic adaptation during inference. From a generalization perspective, ParetoCL can be seen as an objective augmentation approach that learns from different objective combinations of stability and plasticity. Extensive experiments across multiple datasets and settings demonstrate that ParetoCL outperforms state-of-the-art methods and adapts to diverse continual learning scenarios.

Downloads

Published

2025-04-11

How to Cite

Lai, S., Zhao, Z., Zhu, F., Lin, X., Zhang, Q., & Meng, G. (2025). Pareto Continual Learning: Preference-Conditioned Learning and Adaption for Dynamic Stability-Plasticity Trade-off. Proceedings of the AAAI Conference on Artificial Intelligence, 39(17), 18008-18016. https://doi.org/10.1609/aaai.v39i17.33981

Issue

Section

AAAI Technical Track on Machine Learning III