Remember What You Want to Forget: Algorithms for Machine Unlearning
- FedMLMU
We study the problem of forgetting datapoints from a learnt model. In this case, the learner first receives a dataset drawn i.i.d. from an unknown distribution, and outputs a predictor that performs well on unseen samples from that distribution. However, at some point in the future, any training data point can request to be unlearned, thus prompting the learner to modify its output predictor while still ensuring the same accuracy guarantees. In our work, we initiate a rigorous study of machine unlearning in the population setting, where the goal is to maintain performance on the unseen test loss. We then provide unlearning algorithms for convex loss functions. For the setting of convex losses, we provide an unlearning algorithm that can delete up to samples, where is the problem dimension. In comparison, in general, differentially private learningv(which implies unlearning) only guarantees deletion of samples. This shows that unlearning is at least polynomially more efficient than learning privately in terms of dependence on in the deletion capacity.
View on arXiv