ELDER: Enhancing Lifelong Model Editing with Mixture-of-LoRA

Authors

  • Jiaang Li University of Science and Technology of China
  • Quan Wang Beijing University of Posts and Telecommunications
  • Zhongnan Wang University of Science and Technology of China
  • Yongdong Zhang University of Science and Technology of China
  • Zhendong Mao University of Science and Technology of China

DOI:

https://doi.org/10.1609/aaai.v39i23.34622

Abstract

Large language models (LLMs) require model editing to efficiently update specific knowledge within them and avoid factual errors. Most model editing methods are solely designed for single-time use and result in a significant forgetting effect in lifelong editing scenarios, where sequential edits are conducted over time. Previous approaches manage sequential edits by freezing original parameters and discretely allocating new parameters for each knowledge update. However, these methods lack robustness to minor input variations due to the discrete mapping between data and parameters. To overcome this challenge, we propose ELDER, a novel approach to create a continuous association between data and adapters. ELDER integrates multiple LoRAs through a router network and is trained to establish a smooth data-adapter association, thereby enhancing the edit robustness and generalization of semantically equivalent inputs. To ensure inputs containing the same knowledge will be processed by the same LoRAs, we design a novel loss to guide the model link LoRA allocations with edit knowledge. Furthermore, we propose a deferral mechanism to retain the original LLM capabilities post-edit. Extensive experiments on GPT-2 XL and LLaMA2-7B demonstrate that ELDER effectively edits models in the lifelong setting, outperforming eight baselines while exhibiting strong scalability and preserving LLMs' general abilities on downstream tasks.

Published

2025-04-11

How to Cite

Li, J., Wang, Q., Wang, Z., Zhang, Y., & Mao, Z. (2025). ELDER: Enhancing Lifelong Model Editing with Mixture-of-LoRA. Proceedings of the AAAI Conference on Artificial Intelligence, 39(23), 24440–24448. https://doi.org/10.1609/aaai.v39i23.34622

Issue

Section

AAAI Technical Track on Natural Language Processing II