CMT: A Memory Compression Method for Continual Knowledge Learning of Large Language Models

Authors

  • Dongfang Li Harbin Institute of Technology, Shenzhen
  • Zetian Sun Harbin Institute of Technology, Shenzhen
  • Xinshuo Hu Harbin Institute of Technology, Shenzhen
  • Baotian Hu Harbin Institute of Technology, Shenzhen
  • Min Zhang Harbin Institute of Technology, Shenzhen

DOI:

https://doi.org/10.1609/aaai.v39i23.34619

Abstract

Large Language Models (LLMs) need to adapt to the continuous changes in data, tasks, and user preferences. Due to their massive size and the high costs associated with training, LLMs are not suitable for frequent retraining. However, updates are necessary to keep them in sync with rapidly evolving human knowledge. To address these challenges, this paper proposes the Compression Memory Training (CMT) method, an efficient and effective online adaptation framework for LLMs that features robust knowledge retention capabilities. Inspired by human memory mechanisms, CMT compresses and extracts information from new documents to be stored in a memory bank. When answering to queries related to these new documents, the model aggregates these document memories from the memory bank to better answer user questions. The parameters of the LLM itself do not change during training and inference, reducing the risk of catastrophic forgetting. To enhance the encoding, retrieval, and aggregation of memory, we further propose three new general and flexible techniques, including memory-aware objective, self-matching and top-k aggregation. Extensive experiments conducted on three continual learning datasets (i.e., StreamingQA, SQuAD and ArchivalQA) demonstrate that the proposed method improves model adaptability and robustness across multiple base LLMs (e.g., +4.07 EM & +4.19 F1 in StreamingQA with Llama-2-7b).

Downloads

Published

2025-04-11

How to Cite

Li, D., Sun, Z., Hu, X., Hu, B., & Zhang, M. (2025). CMT: A Memory Compression Method for Continual Knowledge Learning of Large Language Models. Proceedings of the AAAI Conference on Artificial Intelligence, 39(23), 24413–24421. https://doi.org/10.1609/aaai.v39i23.34619

Issue

Section

AAAI Technical Track on Natural Language Processing II