Editing Is a Bargaining Game: Balanced Knowledge Editing in Large Language Models

Authors

  • Chenghao Xu Xidian University
  • Jiexi Yan Xidian University
  • Muli Yang A*STAR
  • Fen Fang A*STAR
  • Huilin Chen Xidian University
  • Cheng Deng Xidian University

DOI:

https://doi.org/10.1609/aaai.v40i40.40704

Abstract

Large Language Models (LLMs) are prone to generating incorrect or outdated information, thereby necessitating efficient and precise mechanisms for knowledge updates. Existing knowledge editing approaches, however, often encounter conflicts between two competing objectives: maintaining existing knowledge (preservation) and incorporating new information (editing). During gradient-based optimization, these conflicting objectives can lead to imbalanced update directions, where one gradient dominates, ultimately resulting in suboptimal learning dynamics. To address this challenge, we propose a balanced knowledge editing framework inspired by Nash bargaining theory. Our method guides the optimization process toward a Pareto stationary point, ensuring an equilibrium solution wherein any deviation from the final state would degrade the overall performance with respect to both objectives. This guarantees optimality in preserving prior knowledge while integrating new information. We empirically validate the effectiveness of our approach across a range of evaluation metrics on standard benchmark datasets. Extensive experiments show that our method consistently outperforms state-of-the-art techniques, achieving a superior balance between knowledge preservation and update accuracy.

Downloads

Published

2026-03-14

How to Cite

Xu, C., Yan, J., Yang, M., Fang, F., Chen, H., & Deng, C. (2026). Editing Is a Bargaining Game: Balanced Knowledge Editing in Large Language Models. Proceedings of the AAAI Conference on Artificial Intelligence, 40(40), 34097–34105. https://doi.org/10.1609/aaai.v40i40.40704

Issue

Section

AAAI Technical Track on Natural Language Processing V