ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2211.01916
11
124

Improved Analysis of Score-based Generative Modeling: User-Friendly Bounds under Minimal Smoothness Assumptions

3 November 2022
Hongrui Chen
Holden Lee
Jianfeng Lu
    DiffM
ArXivPDFHTML
Abstract

We give an improved theoretical analysis of score-based generative modeling. Under a score estimate with small L2L^2L2 error (averaged across timesteps), we provide efficient convergence guarantees for any data distribution with second-order moment, by either employing early stopping or assuming smoothness condition on the score function of the data distribution. Our result does not rely on any log-concavity or functional inequality assumption and has a logarithmic dependence on the smoothness. In particular, we show that under only a finite second moment condition, approximating the following in reverse KL divergence in ϵ\epsilonϵ-accuracy can be done in O~(dlog⁡(1/δ)ϵ)\tilde O\left(\frac{d \log (1/\delta)}{\epsilon}\right)O~(ϵdlog(1/δ)​) steps: 1) the variance-δ\deltaδ Gaussian perturbation of any data distribution; 2) data distributions with 1/δ1/\delta1/δ-smooth score functions. Our analysis also provides a quantitative comparison between different discrete approximations and may guide the choice of discretization points in practice.

View on arXiv
Comments on this paper