202

PREM: Privately Answering Statistical Queries with Relative Error

Annual Conference Computational Learning Theory (COLT), 2025
Main:12 Pages
Bibliography:3 Pages
1 Tables
Appendix:11 Pages
Abstract

We introduce PREM\mathsf{PREM} (Private Relative Error Multiplicative weight update), a new framework for generating synthetic data that achieves a relative error guarantee for statistical queries under (ε,δ)(\varepsilon, \delta) differential privacy (DP). Namely, for a domain X{\cal X}, a family F{\cal F} of queries f:X{0,1}f : {\cal X} \to \{0, 1\}, and ζ>0\zeta > 0, our framework yields a mechanism that on input dataset DXnD \in {\cal X}^n outputs a synthetic dataset D^Xn\widehat{D} \in {\cal X}^n such that all statistical queries in F{\cal F} on DD, namely xDf(x)\sum_{x \in D} f(x) for fFf \in {\cal F}, are within a 1±ζ1 \pm \zeta multiplicative factor of the corresponding value on D^\widehat{D} up to an additive error that is polynomial in logF\log |{\cal F}|, logX\log |{\cal X}|, logn\log n, log(1/δ)\log(1/\delta), 1/ε1/\varepsilon, and 1/ζ1/\zeta. In contrast, any (ε,δ)(\varepsilon, \delta)-DP mechanism is known to require worst-case additive error that is polynomial in at least one of n,Fn, |{\cal F}|, or X|{\cal X}|. We complement our algorithm with nearly matching lower bounds.

View on arXiv
Comments on this paper