31

Less Noise, Same Certificate: Retain Sensitivity for Unlearning

Carolin Heinzler
Kasra Malihi
Amartya Sanyal
Main:10 Pages
3 Figures
Bibliography:4 Pages
5 Tables
Appendix:12 Pages
Abstract

Certified machine unlearning aims to provably remove the influence of a deletion set UU from a model trained on a dataset SS, by producing an unlearned output that is statistically indistinguishable from retraining on the retain set R:=SUR:=S\setminus U. Many existing certified unlearning methods adapt techniques from Differential Privacy (DP) and add noise calibrated to global sensitivity, i.e., the worst-case output change over all adjacent datasets. We show that this DP-style calibration is often overly conservative for unlearning, based on a key observation: certified unlearning, by definition, does not require protecting the privacy of the retained data RR. Motivated by this distinction, we define retain sensitivity as the worst-case output change over deletions UU while keeping RR fixed. While insufficient for DP, retain sensitivity is exactly sufficient for unlearning, allowing for the same certificates with less noise. We validate these reductions in noise theoretically and empirically across several problems, including the weight of minimum spanning trees, PCA, and ERM. Finally, we refine the analysis of two widely used certified unlearning algorithms through the lens of retain sensitivity, leveraging the regularity induced by RR to further reduce noise and improve utility.

View on arXiv
Comments on this paper