27
2

Rewind-to-Delete: Certified Machine Unlearning for Nonconvex Functions

Abstract

Machine unlearning algorithms aim to efficiently remove data from a model without retraining it from scratch, in order to remove corrupted or outdated data or respect a user's ``right to be forgotten." Certified machine unlearning is a strong theoretical guarantee based on differential privacy that quantifies the extent to which an algorithm erases data from the model weights. In contrast to existing works in certified unlearning for convex or strongly convex loss functions, or nonconvex objectives with limiting assumptions, we propose the first, first-order, black-box (i.e., can be applied to models pretrained with vanilla gradient descent) algorithm for unlearning on general nonconvex loss functions, which unlearns by ``rewinding" to an earlier step during the learning process before performing gradient descent on the loss function of the retained data points. We prove (ϵ,δ)(\epsilon, \delta) certified unlearning and performance guarantees that establish the privacy-utility-complexity tradeoff of our algorithm, and we prove generalization guarantees for nonconvex functions that satisfy the Polyak-Lojasiewicz inequality. Finally, we implement our algorithm under a new experimental framework that more accurately reflects real-world use cases for preserving user privacy.

View on arXiv
@article{mu2025_2409.09778,
  title={ Rewind-to-Delete: Certified Machine Unlearning for Nonconvex Functions },
  author={ Siqiao Mu and Diego Klabjan },
  journal={arXiv preprint arXiv:2409.09778},
  year={ 2025 }
}
Comments on this paper