Quantifying Classifier Utility under Local Differential Privacy
Main:11 Pages
16 Figures
Bibliography:3 Pages
3 Tables
Appendix:8 Pages
Abstract
Local differential privacy (LDP) offers rigorous, quantifiable privacy guarantees for personal data by introducing perturbations at the data source. Understanding how these perturbations affect classifier utility is crucial for both designers and users. However, a general theoretical framework for quantifying this impact is lacking and also challenging, especially for complex or black-box classifiers.
View on arXivComments on this paper
