Injection, Attack and Erasure: Revocable Backdoor Attacks via Machine Unlearning

Authors

  • Baogang Song Wuhan University of Technology
  • Dongdong Zhao Wuhan University of Technology
  • Jianwen Xiang Wuhan University of Technology
  • Qiben Xu Wuhan University of Technology
  • Zizhuo Yu Wuhan University of Technology

DOI:

https://doi.org/10.1609/aaai.v40i30.39747

Abstract

Backdoor attacks pose a persistent security risk to deep neural networks (DNNs) due to their stealth and durability. While recent research has explored leveraging model unlearning mechanisms to enhance backdoor concealment, existing attack strategies still leave persistent traces that may be detected through static analysis. In this work, we introduce the first paradigm of revocable backdoor attacks, where the backdoor can be proactively and thoroughly removed after the attack objective is achieved. We formulate the trigger optimization in revocable backdoor attacks as a bilevel optimization problem: by simulating both backdoor injection and unlearning processes, the trigger generator is optimized to achieve a high attack success rate (ASR) while ensuring that the backdoor can be easily erased through unlearning. To mitigate the optimization conflict between injection and removal objectives, we employ a deterministic partition of poisoning and unlearning samples to reduce sampling-induced variance, and further apply the Projected Conflicting Gradient (PCGrad) technique to resolve the remaining gradient conflicts. Experiments on CIFAR-10 and ImageNet demonstrate that our method maintains ASR comparable to state-of-the-art backdoor attacks, while enabling effective removal of backdoor behavior after unlearning. This work opens a new direction for backdoor attack research and presents new challenges for the security of machine learning systems.

Downloads

Published

2026-03-14

How to Cite

Song, B., Zhao, D., Xiang, J., Xu, Q., & Yu, Z. (2026). Injection, Attack and Erasure: Revocable Backdoor Attacks via Machine Unlearning. Proceedings of the AAAI Conference on Artificial Intelligence, 40(30), 25518–25526. https://doi.org/10.1609/aaai.v40i30.39747

Issue

Section

AAAI Technical Track on Machine Learning VII