106
0

What Exactly Does Guidance Do in Masked Discrete Diffusion Models

Main:11 Pages
23 Figures
Bibliography:3 Pages
1 Tables
Appendix:25 Pages
Abstract

We study masked discrete diffusion models with classifier-free guidance (CFG). Assuming no score error nor discretization error, we derive an explicit solution to the guided reverse dynamics, so that how guidance influences the sampling behavior can be precisely characterized. When the full data distribution is a mixture over classes and the goal is to sample from a specific class, guidance amplifies class-specific regions while suppresses regions shared with other classes. This effect depends on the guidance strength ww and induces distinct covariance structures in the sampled distribution. Notably, we observe quantitatively different behaviors in 11D and 22D. We also show that for large ww, the decay rate of the total variation (TV\mathrm{TV}) along the reverse dynamics is double-exponential in ww for both 11D and 22D. These findings highlight the role of guidance, not just in shaping the output distribution, but also in controlling the dynamics of the sampling trajectory. Our theoretical analysis is supported by experiments that illustrate the geometric effects of guidance and its impact on convergence.

View on arXiv
@article{ye2025_2506.10971,
  title={ What Exactly Does Guidance Do in Masked Discrete Diffusion Models },
  author={ He Ye and Rojas Kevin and Tao Molei },
  journal={arXiv preprint arXiv:2506.10971},
  year={ 2025 }
}
Comments on this paper