Iterative Tilting for Diffusion Fine-Tuning
- DiffM
Main:11 Pages
2 Figures
Bibliography:3 Pages
1 Tables
Abstract
We introduce iterative tilting, a gradient-free method for fine-tuning diffusion models toward reward-tilted distributions. The method decomposes a large reward tilt into sequential smaller tilts, each admitting a tractable score update via first-order Taylor expansion. This requires only forward evaluations of the reward function and avoids backpropagating through sampling chains. We validate on a two-dimensional Gaussian mixture with linear reward, where the exact tilted distribution is available in closed form.
View on arXivComments on this paper
