Learning to Rewind via Iterative Prediction of Past Weights for Practical Unlearning

Authors

  • Jinhyeok Jang ETRI KAIST
  • Jaehong Kim ETRI
  • Chan-Hyun Youn KAIST

DOI:

https://doi.org/10.1609/aaai.v39i25.34822

Abstract

In artificial intelligence (AI), many legal conflicts have arisen, especially concerning privacy and copyright associated with training data. When an AI model's training data incurs privacy concerns, it becomes imperative to develop a new model devoid of influences from such contentious data. However, retraining from scratch is often not viable due to the extensive data requirements and heavy computational costs. Machine unlearning presents a promising solution by enabling the selective erasure of specific knowledge from models. Despite its potential, many existing approaches in machine unlearning are based on scenarios that are either impractical or could lead to unintended degradation of model performance. We utilize the concept of weight prediction to approximate the less-learned weights based on observations about further training. By repetition of 1) finetuning on specific data and 2) weight prediction, our work gradually eliminates knowledge about the specific data. We verify its ability to eliminate side effects caused by problematic data and show its effectiveness across various architectures, datasets, and tasks.

Downloads

Published

2025-04-11

How to Cite

Jang, J., Kim, J., & Youn, C.-H. (2025). Learning to Rewind via Iterative Prediction of Past Weights for Practical Unlearning. Proceedings of the AAAI Conference on Artificial Intelligence, 39(25), 26248–26255. https://doi.org/10.1609/aaai.v39i25.34822

Issue

Section

AAAI Technical Track on Philosophy and Ethics of AI