99
v1v2 (latest)

Identifying Memorization of Diffusion Models through pp-Laplace Analysis: Estimators, Bounds and Applications

Scale Space and Variational Methods in Computer Vision (SSVM), 2025
Main:14 Pages
5 Figures
Bibliography:5 Pages
1 Tables
Abstract

Diffusion models, today's leading image generative models, estimate the score function, i.e. the gradient of the log probability of (perturbed) data samples, without direct access to the underlying probability distribution. This work investigates whether the estimated score function can be leveraged to compute higher-order differentials, namely the p-Laplace operators. We show that these operators can be employed to identify memorized training data. We propose a numerical p-Laplace approximation based on the learned score functions, showing its effectiveness in identifying key features of the probability landscape. Furthermore, theoretical error-bounds to these estimators are proven and demonstrated numerically. We analyze the structured case of Gaussian mixture models, and demonstrate that the results carry-over to text-conditioned image generative models (text-to-image), where memorization identification based on the p-Laplace operator is performed for the first time, showing its advantage on 500 memorized prompts (\sim3000 generated images) in a post-generation regime, especially when the conditioning text is unavailable.

View on arXiv
Comments on this paper