Differentially Private Learning with Small Public Data

Authors

  • Jun Wang Nanjing University
  • Zhi-Hua Zhou Nanjing University

DOI:

https://doi.org/10.1609/aaai.v34i04.6088

Abstract

Differentially private learning tackles tasks where the data are private and the learning process is subject to differential privacy requirements. In real applications, however, some public data are generally available in addition to private data, and it is interesting to consider how to exploit them. In this paper, we study a common situation where a small amount of public data can be used when solving the Empirical Risk Minimization problem over a private database. Specifically, we propose Private-Public Stochastic Gradient Descent, which utilizes such public information to adjust parameters in differentially private stochastic gradient descent and fine-tunes the final result with model reuse. Our method keeps differential privacy for the private database, and empirical study validates its superiority compared with existing approaches.

Downloads

Published

2020-04-03

How to Cite

Wang, J., & Zhou, Z.-H. (2020). Differentially Private Learning with Small Public Data. Proceedings of the AAAI Conference on Artificial Intelligence, 34(04), 6219-6226. https://doi.org/10.1609/aaai.v34i04.6088

Issue

Section

AAAI Technical Track: Machine Learning