Task-Free Continual Generation and Representation Learning via Dynamic Expansionable Memory Cluster

Authors

  • Fei Ye Department of Computer Science, University of York, York YO10 5GH, UK Machine Learning Department, Mohamed bin Zayed University of Artificial Intelligence, Abu Dhabi, UAE
  • Adrian G. Bors Department of Computer Science, University of York, York YO10 5GH, UK Machine Learning Department, Mohamed bin Zayed University of Artificial Intelligence, Abu Dhabi, UAE

DOI:

https://doi.org/10.1609/aaai.v38i15.29582

Keywords:

ML: Online Learning & Bandits, ML: Classification and Regression, ML: Deep Generative Models & Autoencoders

Abstract

Human brains can continually acquire and learn new skills and knowledge over time from a dynamically changing environment without forgetting previously learnt information. Such a capacity can selectively transfer some important and recently seen information to the persistent knowledge regions of the brain. Inspired by this intuition, we propose a new memory-based approach for image reconstruction and generation in continual learning, consisting of a temporary and evolving memory, with two different storage strategies, corresponding to the temporary and permanent memorisation. The temporary memory aims to preserve up-to-date information while the evolving memory can dynamically increase its capacity in order to preserve permanent knowledge information. This is achieved by the proposed memory expansion mechanism that selectively transfers those data samples deemed as important from the temporary memory to new clusters defined within the evolved memory according to an information novelty criterion. Such a mechanism promotes the knowledge diversity among clusters in the evolved memory, resulting in capturing more diverse information by using a compact memory capacity. Furthermore, we propose a two-step optimization strategy for training a Variational Autoencoder (VAE) to implement generation and representation learning tasks, which updates the generator and inference models separately using two optimisation paths. This approach leads to a better trade-off between generation and reconstruction performance. We show empirically and theoretically that the proposed approach can learn meaningful latent representations while generating diverse images from different domains. The source code and supplementary material (SM) are available at https://github.com/dtuzi123/DEMC.

Published

2024-03-24

How to Cite

Ye, F., & Bors, A. G. (2024). Task-Free Continual Generation and Representation Learning via Dynamic Expansionable Memory Cluster. Proceedings of the AAAI Conference on Artificial Intelligence, 38(15), 16451–16459. https://doi.org/10.1609/aaai.v38i15.29582

Issue

Section

AAAI Technical Track on Machine Learning VI