DP-AdamBC: Your DP-Adam Is Actually DP-SGD (Unless You Apply Bias Correction)

Authors

  • Qiaoyue Tang University of British Columbia
  • Frederick Shpilevskiy University of British Columbia
  • Mathias Lécuyer University of British Columbia

DOI:

https://doi.org/10.1609/aaai.v38i14.29451

Keywords:

ML: Privacy

Abstract

The Adam optimizer is a popular choice in contemporary deep learning due to its strong empirical performance. However we observe that in privacy sensitive scenarios, the traditional use of Differential Privacy (DP) with the Adam optimizer leads to sub-optimal performance on several tasks. We find that this performance degradation is due to a DP bias in Adam's second moment estimator, introduced by the addition of independent noise in the gradient computation to enforce DP guarantees. This DP bias leads to a different scaling for low variance parameter updates, that is inconsistent with the behavior of non-private Adam, and Adam's sign descent interpretation. We propose the DP-AdamBC optimization algorithm, which corrects for the bias in the second moment estimation and retrieves the expected behaviour of Adam. Empirically, DP-AdamBC significantly improves the optimization performance of DP-Adam by up to 3.5% in final accuracy in image, text, and graph node classification tasks.

Published

2024-03-24

How to Cite

Tang, Q., Shpilevskiy, F., & Lécuyer, M. (2024). DP-AdamBC: Your DP-Adam Is Actually DP-SGD (Unless You Apply Bias Correction). Proceedings of the AAAI Conference on Artificial Intelligence, 38(14), 15276-15283. https://doi.org/10.1609/aaai.v38i14.29451

Issue

Section

AAAI Technical Track on Machine Learning V