Differential Privacy Stochastic Gradient Descent with Adaptive Privacy Budget Allocation

Author(s):  
Yun Xie ◽  
Peng Li ◽  
Chao Wu ◽  
Qiuling Wu
2020 ◽  
Vol 34 (04) ◽  
pp. 6219-6226
Author(s):  
Jun Wang ◽  
Zhi-Hua Zhou

Differentially private learning tackles tasks where the data are private and the learning process is subject to differential privacy requirements. In real applications, however, some public data are generally available in addition to private data, and it is interesting to consider how to exploit them. In this paper, we study a common situation where a small amount of public data can be used when solving the Empirical Risk Minimization problem over a private database. Specifically, we propose Private-Public Stochastic Gradient Descent, which utilizes such public information to adjust parameters in differentially private stochastic gradient descent and fine-tunes the final result with model reuse. Our method keeps differential privacy for the private database, and empirical study validates its superiority compared with existing approaches.


Sign in / Sign up

Export Citation Format

Share Document