Dropout Is NOT All You Need to Prevent Gradient Leakage

Authors

  • Daniel Scheliga Technische Universität Ilmenau, Ilmenau, Germany
  • Patrick Maeder Technische Universität Ilmenau, Germany Friedrich Schiller Universität Jena, Germany
  • Marco Seeland Technische Universität Ilmenau, Germany

DOI:

https://doi.org/10.1609/aaai.v37i8.26163

Keywords:

ML: Privacy-Aware ML, CV: Bias, Fairness & Privacy, ML: Distributed Machine Learning & Federated Learning

Abstract

Gradient inversion attacks on federated learning systems reconstruct client training data from exchanged gradient information. To defend against such attacks, a variety of defense mechanisms were proposed. However, they usually lead to an unacceptable trade-off between privacy and model utility. Recent observations suggest that dropout could mitigate gradient leakage and improve model utility if added to neural networks. Unfortunately, this phenomenon has not been systematically researched yet. In this work, we thoroughly analyze the effect of dropout on iterative gradient inversion attacks. We find that state of the art attacks are not able to reconstruct the client data due to the stochasticity induced by dropout during model training. Nonetheless, we argue that dropout does not offer reliable protection if the dropout induced stochasticity is adequately modeled during attack optimization. Consequently, we propose a novel Dropout Inversion Attack (DIA) that jointly optimizes for client data and dropout masks to approximate the stochastic client model. We conduct an extensive systematic evaluation of our attack on four seminal model architectures and three image classification datasets of increasing complexity. We find that our proposed attack bypasses the protection seemingly induced by dropout and reconstructs client data with high fidelity. Our work demonstrates that privacy inducing changes to model architectures alone cannot be assumed to reliably protect from gradient leakage and therefore should be combined with complementary defense mechanisms.

Downloads

Published

2023-06-26

How to Cite

Scheliga, D., Maeder, P., & Seeland, M. (2023). Dropout Is NOT All You Need to Prevent Gradient Leakage. Proceedings of the AAAI Conference on Artificial Intelligence, 37(8), 9733-9741. https://doi.org/10.1609/aaai.v37i8.26163

Issue

Section

AAAI Technical Track on Machine Learning III