578

Bringing Differential Private SGD to Practice: On the Independence of Gaussian Noise and the Number of Training Rounds

International Conference on Machine Learning (ICML), 2021
Abstract

In DP-SGD each round communicates a local SGD update which leaks some new information about the underlying local data set to the outside world. In order to provide privacy, Gaussian noise with standard deviation σ\sigma is added to local SGD updates after performing a clipping operation. We show that for attaining (ϵ,δ)(\epsilon,\delta)-differential privacy σ\sigma can be chosen equal to 2(ϵ+ln(1/δ))/ϵ\sqrt{2(\epsilon +\ln(1/\delta))/\epsilon} for ϵ=Ω(T/N2)\epsilon=\Omega(T/N^2), where TT is the total number of rounds and NN is equal to the size of the local data set. In many existing machine learning problems, NN is always large and T=O(N)T=O(N). Hence, σ\sigma becomes "independent" of any T=O(N)T=O(N) choice with ϵ=Ω(1/N)\epsilon=\Omega(1/N). This means that our σ\sigma only depends on NN rather than TT. As shown in our paper, this differential privacy characterization allows one to {\it a-priori} select parameters of DP-SGD based on a fixed privacy budget (in terms of ϵ\epsilon and δ\delta) in such a way to optimize the anticipated utility (test accuracy) the most. This ability of planning ahead together with σ\sigma's independence of TT (which allows local gradient computations to be split among as many rounds as needed, even for large TT as usually happens in practice) leads to a {\it proactive DP-SGD algorithm} that allows a client to balance its privacy budget with the accuracy of the learned global model based on local test data. We notice that the current state-of-the art differential privacy accountant method based on ff-DP has a closed form for computing the privacy loss for DP-SGD. However, due to its interpretation complexity, it cannot be used in a simple way to plan ahead. Instead, accountant methods are only used for keeping track of how privacy budget has been spent (after the fact).

View on arXiv
Comments on this paper