34
16

Differentially Private Generalized Linear Models Revisited

Abstract

We study the problem of (ϵ,δ)(\epsilon,\delta)-differentially private learning of linear predictors with convex losses. We provide results for two subclasses of loss functions. The first case is when the loss is smooth and non-negative but not necessarily Lipschitz (such as the squared loss). For this case, we establish an upper bound on the excess population risk of O~(wn+min{w2(nϵ)2/3,dw2nϵ})\tilde{O}\left(\frac{\Vert w^*\Vert}{\sqrt{n}} + \min\left\{\frac{\Vert w^* \Vert^2}{(n\epsilon)^{2/3}},\frac{\sqrt{d}\Vert w^*\Vert^2}{n\epsilon}\right\}\right), where nn is the number of samples, dd is the dimension of the problem, and ww^* is the minimizer of the population risk. Apart from the dependence on w\Vert w^\ast\Vert, our bound is essentially tight in all parameters. In particular, we show a lower bound of Ω~(1n+min{w4/3(nϵ)2/3,dwnϵ})\tilde{\Omega}\left(\frac{1}{\sqrt{n}} + {\min\left\{\frac{\Vert w^*\Vert^{4/3}}{(n\epsilon)^{2/3}}, \frac{\sqrt{d}\Vert w^*\Vert}{n\epsilon}\right\}}\right). We also revisit the previously studied case of Lipschitz losses [SSTT20]. For this case, we close the gap in the existing work and show that the optimal rate is (up to log factors) Θ(wn+min{wnϵ,rankwnϵ})\Theta\left(\frac{\Vert w^*\Vert}{\sqrt{n}} + \min\left\{\frac{\Vert w^*\Vert}{\sqrt{n\epsilon}},\frac{\sqrt{\text{rank}}\Vert w^*\Vert}{n\epsilon}\right\}\right), where rank\text{rank} is the rank of the design matrix. This improves over existing work in the high privacy regime. Finally, our algorithms involve a private model selection approach that we develop to enable attaining the stated rates without a-priori knowledge of w\Vert w^*\Vert.

View on arXiv
Comments on this paper