38
1

Differential Private Stochastic Optimization with Heavy-tailed Data: Towards Optimal Rates

Abstract

We study convex optimization problems under differential privacy (DP). With heavy-tailed gradients, existing works achieve suboptimal rates. The main obstacle is that existing gradient estimators have suboptimal tail properties, resulting in a superfluous factor of dd in the union bound. In this paper, we explore algorithms achieving optimal rates of DP optimization with heavy-tailed gradients. Our first method is a simple clipping approach. Under bounded pp-th order moments of gradients, with nn samples, it achieves O~(d/n+d(d/nϵ)11/p)\tilde{O}(\sqrt{d/n}+\sqrt{d}(\sqrt{d}/n\epsilon)^{1-1/p}) population risk with ϵ1/d\epsilon\leq 1/\sqrt{d}. We then propose an iterative updating method, which is more complex but achieves this rate for all ϵ1\epsilon\leq 1. The results significantly improve over existing methods. Such improvement relies on a careful treatment of the tail behavior of gradient estimators. Our results match the minimax lower bound in \cite{kamath2022improved}, indicating that the theoretical limit of stochastic convex optimization under DP is achievable.

View on arXiv
Comments on this paper