Beyond Binary Erasure: Soft-Weighted Unlearning for Fairness and Robustness
DOI:
https://doi.org/10.1609/aaai.v40i29.39681Abstract
Machine unlearning, as a post-hoc processing technique, has gained widespread adoption in addressing challenges like bias mitigation and robustness enhancement. However, existing non-privacy unlearning-based solutions persist in using a binary data removal framework designed for privacy-driven motivation, even when repurposed for fairness or robustness improvements. This leads to significant utility loss, a phenomenon known as “over-unlearning”. While over-unlearning has been largely described in many studies as primarily causing utility degradation, we investigate deeper insights in this work through counterfactual leave-one-out analysis. Based on insights, we introduce a soft weighting strategy that assigns tailored weights to each sample by solving a convex quadratic programming problem analytically, which enables fine-grained model adjustments to address the over-unlearning. We demonstrate that the proposed soft-weighted scheme can be seamlessly integrated into most existing unlearning algorithms. Extensive experiments show that in fairness- and robustness-driven tasks, the soft-weighted scheme significantly outperforms hard-weighted schemes in fairness/robustness metrics and alleviates the decline in utility metric, thereby enhancing unlearning algorithm as an effective correction solution.Published
2026-03-14
How to Cite
Qiao, X., Ding, N., Cheng, Y., & Zhang, M. (2026). Beyond Binary Erasure: Soft-Weighted Unlearning for Fairness and Robustness. Proceedings of the AAAI Conference on Artificial Intelligence, 40(29), 24936-24944. https://doi.org/10.1609/aaai.v40i29.39681
Issue
Section
AAAI Technical Track on Machine Learning VI