65
52

Private Convex Optimization via Exponential Mechanism

Abstract

In this paper, we study private optimization problems for non-smooth convex functions F(x)=Eifi(x)F(x)=\mathbb{E}_i f_i(x) on Rd\mathbb{R}^d. We show that modifying the exponential mechanism by adding an 22\ell_2^2 regularizer to F(x)F(x) and sampling from π(x)exp(k(F(x)+μx22/2))\pi(x)\propto \exp(-k(F(x)+\mu\|x\|_2^2/2)) recovers both the known optimal empirical risk and population loss under (ϵ,δ)(\epsilon,\delta)-DP. Furthermore, we show how to implement this mechanism using O~(nmin(d,n))\widetilde{O}(n \min(d, n)) queries to fi(x)f_i(x) for the DP-SCO where nn is the number of samples/users and dd is the ambient dimension. We also give a (nearly) matching lower bound Ω~(nmin(d,n))\widetilde{\Omega}(n \min(d, n)) on the number of evaluation queries. Our results utilize the following tools that are of independent interest: (1) We prove Gaussian Differential Privacy (GDP) of the exponential mechanism if the loss function is strongly convex and the perturbation is Lipschitz. Our privacy bound is \emph{optimal} as it includes the privacy of Gaussian mechanism as a special case and is proved using the isoperimetric inequality for strongly log-concave measures. (2) We show how to sample from exp(F(x)μx22/2)\exp(-F(x)-\mu \|x\|^2_2/2) for GG-Lipschitz FF with η\eta error in total variation (TV) distance using O~((G2/μ)log2(d/η))\widetilde{O}((G^2/\mu) \log^2(d/\eta)) unbiased queries to F(x)F(x). This is the first sampler whose query complexity has \emph{polylogarithmic dependence} on both dimension dd and accuracy η\eta.

View on arXiv
Comments on this paper