297
v1v2 (latest)

Optimal Convergence Analysis of DDPM for General Distributions

Main:29 Pages
1 Figures
Bibliography:3 Pages
Abstract

Score-based diffusion models have achieved remarkable empirical success in generating high-quality samples from target data distributions. Among them, the Denoising Diffusion Probabilistic Model (DDPM) is one of the most widely used samplers, generating samples via estimated score functions. Despite its empirical success, a tight theoretical understanding of DDPM -- especially its convergence properties -- remains limited.In this paper, we provide a refined convergence analysis of the DDPM sampler and establish near-optimal convergence rates under general distributional assumptions. Specifically, we introduce a relaxed smoothness condition parameterized by a constant LL, which is small for many practical distributions (e.g., Gaussian mixture models). We prove that the DDPM sampler with accurate score estimates achieves a convergence rate of O~(dmin{d,L2}T2) in Kullback-Leibler divergence,\widetilde{O}\left(\frac{d\min\{d,L^2\}}{T^2}\right)~\text{in Kullback-Leibler divergence}, where dd is the data dimension, TT is the number of iterations, and O~\widetilde{O} hides polylogarithmic factors in TT. This result substantially improves upon the best-known d2/T2d^2/T^2 rate when L<dL < \sqrt{d}. By establishing a matching lower bound, we show that our convergence analysis is tight for a wide array of target distributions. Moreover, it reveals that DDPM and DDIM share the same dependence on dd, raising an interesting question of why DDIM often appears empirically faster.

View on arXiv
Comments on this paper