37

An explicit formulation of the learned noise predictor εθ(xt,t)ε_θ({\bf x}_t, t) via the forward-process noise εtε_{t} in denoising diffusion probabilistic models (DDPMs)

KiHyun Yun
Main:3 Pages
Bibliography:1 Pages
Appendix:1 Pages
Abstract

In denoising diffusion probabilistic models (DDPMs), the learned noise predictor $ \epsilon_{\theta} ( {\bf x}_t , t)$ is trained to approximate the forward-process noise ϵt\epsilon_t. The equality xtlogq(xt)=11αˉtϵθ(xt,t)\nabla_{{\bf x}_t} \log q({\bf x}_t) = -\frac 1 {\sqrt {1- {\bar \alpha}_t} } \epsilon_{\theta} ( {\bf x}_t , t) plays a fundamental role in both theoretical analyses and algorithmic design, and thus is frequently employed across diffusion-based generative models. In this paper, an explicit formulation of $ \epsilon_{\theta} ( {\bf x}_t , t)$ in terms of the forward-process noise ϵt\epsilon_t is derived. This result show how the forward-process noise ϵt\epsilon_t contributes to the learned predictor $ \epsilon_{\theta} ( {\bf x}_t , t)$. Furthermore, based on this formulation, we present a novel and mathematically rigorous proof of the fundamental equality above, clarifying its origin and providing new theoretical insight into the structure of diffusion models.

View on arXiv
Comments on this paper