Differentially Private Kernel Density Estimation

We introduce a refined differentially private (DP) data structure for kernel density estimation (KDE), offering not only improved privacy-utility tradeoff but also better efficiency over prior results. Specifically, we study the mathematical problem: given a similarity function (or DP KDE) and a private dataset , our goal is to preprocess so that for any query , we approximate in a differentially private fashion. The best previous algorithm for is the node-contaminated balanced binary tree by [Backurs, Lin, Mahabadi, Silwal, and Tarnawski, ICLR 2024]. Their algorithm requires space and time for preprocessing with . For any query point, the query time is , with an error guarantee of -approximation and .In this paper, we improve the best previous result [Backurs, Lin, Mahabadi, Silwal, and Tarnawski, ICLR 2024] in three aspects:- We reduce query time by a factor of .- We improve the approximation ratio from to 1.- We reduce the error dependence by a factor of .From a technical perspective, our method of constructing the search tree differs from previous work [Backurs, Lin, Mahabadi, Silwal, and Tarnawski, ICLR 2024]. In prior work, for each query, the answer is split into numbers, each derived from the summation of values in interval tree countings. In contrast, we construct the tree differently, splitting the answer into numbers, where each is a smart combination of two distance values, two counting values, and itself. We believe our tree structure may be of independent interest.
View on arXiv@article{liu2025_2409.01688, title={ Differentially Private Kernel Density Estimation }, author={ Erzhi Liu and Jerry Yao-Chieh Hu and Alex Reneau and Zhao Song and Han Liu }, journal={arXiv preprint arXiv:2409.01688}, year={ 2025 } }