PLAN: Variance-Aware Private Mean Estimation

Differentially private mean estimation is an important building block in privacy-preserving algorithms for data analysis and machine learning. Though the trade-off between privacy and utility is well understood in the worst case, many datasets exhibit structure that could potentially be exploited to yield better algorithms. In this paper we present (PLAN), a family of differentially private algorithms for mean estimation in the setting where inputs are independently sampled from a distribution over , with coordinate-wise standard deviations . Similar to mean estimation under Mahalanobis distance, PLAN tailors the shape of the noise to the shape of the data, but unlike previous algorithms the privacy budget is spent non-uniformly over the coordinates. Under a concentration assumption on , we show how to exploit skew in the vector , obtaining a (zero-concentrated) differentially private mean estimate with error proportional to . Previous work has either not taken into account, or measured error in Mahalanobis distance in both cases resulting in error proportional to , which can be up to a factor larger. To verify the effectiveness of PLAN, we empirically evaluate accuracy on both synthetic and real world data.
View on arXiv