Sample-Optimal Locally Private Hypothesis Selection and the Provable Benefits of Interactivity

We study the problem of hypothesis selection under the constraint of local differential privacy. Given a class of distributions and a set of i.i.d. samples from an unknown distribution , the goal of hypothesis selection is to pick a distribution whose total variation distance to is comparable with the best distribution in (with high probability). We devise an -locally-differentially-private (-LDP) algorithm that uses samples to guarantee that with high probability. This sample complexity is optimal for , matching the lower bound of Gopi et al. (2020). All previously known algorithms for this problem required samples to work. Moreover, our result demonstrates the power of interaction for -LDP hypothesis selection. Namely, it breaks the known lower bound of for the sample complexity of non-interactive hypothesis selection. Our algorithm breaks this barrier using only rounds of interaction. To prove our results, we define the notion of \emph{critical queries} for a Statistical Query Algorithm (SQA) which may be of independent interest. Informally, an SQA is said to use a small number of critical queries if its success relies on the accuracy of only a small number of queries it asks. We then design an LDP algorithm that uses a smaller number of critical queries.
View on arXiv