Lifelong Generative Modelling Using Dynamic Expansion Graph Model

Authors

  • Fei Ye University of York
  • Adrian G. Bors University of York

DOI:

https://doi.org/10.1609/aaai.v36i8.20867

Keywords:

Machine Learning (ML), Knowledge Representation And Reasoning (KRR)

Abstract

Variational Autoencoders (VAEs) suffer from degenerated performance, when learning several successive tasks. This is caused by catastrophic forgetting. In order to address the knowledge loss, VAEs are using either Generative Replay (GR) mechanisms or Expanding Network Architectures (ENA). In this paper we study the forgetting behaviour of VAEs using a joint GR and ENA methodology, by deriving an upper bound on the negative marginal log-likelihood. This theoretical analysis provides new insights into how VAEs forget the previously learnt knowledge during lifelong learning. The analysis indicates the best performance achieved when considering model mixtures, under the ENA framework, where there are no restrictions on the number of components. However, an ENA-based approach may require an excessive number of parameters. This motivates us to propose a novel Dynamic Expansion Graph Model (DEGM). DEGM expands its architecture, according to the novelty associated with each new database, when compared to the information already learnt by the network from previous tasks. DEGM training optimizes knowledge structuring, characterizing the joint probabilistic representations corresponding to the past and more recently learned tasks. We demonstrate that DEGM guarantees optimal performance for each task while also minimizing the required number of parameters.

Downloads

Published

2022-06-28

How to Cite

Ye, F., & Bors, A. G. (2022). Lifelong Generative Modelling Using Dynamic Expansion Graph Model. Proceedings of the AAAI Conference on Artificial Intelligence, 36(8), 8857-8865. https://doi.org/10.1609/aaai.v36i8.20867

Issue

Section

AAAI Technical Track on Machine Learning III