Beyond Memorization: Gradient Projection Enables Selective Learning in Diffusion Models
Divya Kothandaraman
Jaclyn Pytlarz
- DiffMAAML
Main:16 Pages
5 Figures
Bibliography:4 Pages
3 Tables
Abstract
Memorization in large-scale text-to-image diffusion models poses significant security and intellectual property risks, enabling adversarial attribute extraction and the unauthorized reproduction of sensitive or proprietary features. While conventional dememorization techniques, such as regularization and data filtering, limit overfitting to specific training examples, they fail to systematically prevent the internalization of prohibited concept-level features. Simply discarding all images containing a sensitive feature wastes invaluable training data, necessitating a method for selective unlearning at the concept level.
View on arXivComments on this paper
