527

Bringing Differential Private SGD to Practice: On the Independence of Gaussian Noise and the Number of Training Rounds

International Conference on Machine Learning (ICML), 2021
Abstract

Different from existing Differential Privacy (DP) accountants, we introduce pro-active DP. Existing DP accountants keep track of how privacy budget has been spent while pro-active DP is a scheme that allows one to {\it a-priori} select parameters of DP-SGD based on a fixed privacy budget (in terms of ϵ\epsilon and δ\delta) in such a way to optimize the anticipated utility (test accuracy) the most. To implement this idea, we show how to convert the classical DP moment accountant to a pro-active DP by exploiting the fact that it has a simple close form for computing spent privacy budget for a given interaction round. The DP moment accountant is introduced in context of DP-SGD and has the following property which is the key ingredient to build pro-active DP. In DP-SGD each round communicates a local SGD update which leaks some new information about the underlying local data set to the outside world. In order to provide privacy, Gaussian noise with standard deviation σ\sigma is added to local SGD updates after performing a clipping operation and normalizing with the clipping constant. We show that for attaining (ϵ,δ)(\epsilon,\delta)-differential privacy σ\sigma can be chosen equal to 2(ϵ+ln(1/δ))/ϵ\sqrt{2(\epsilon +\ln(1/\delta))/\epsilon} for ϵ=Ω(T/N2)\epsilon=\Omega(T/N^2), where TT is the total number of rounds and NN is equal to the size of the local data set. In many existing machine learning problems, NN is always large and T=O(N)T=O(N). Hence, σ\sigma becomes "independent" of any T=O(N)T=O(N) choice with ϵ=Ω(1/N)\epsilon=\Omega(1/N). This means that our {\em σ\sigma only depends on NN rather than TT}. We show how this differential privacy characterization allows us to convert DP moment accountant to a pro-active DP.

View on arXiv
Comments on this paper