v1v2 (latest)
Bellman Diffusion Models
- DiffM
Main:7 Pages
Bibliography:2 Pages
5 Tables
Appendix:8 Pages
Abstract
Diffusion models have seen tremendous success as generative architectures. Recently, they have been shown to be effective at modelling policies for offline reinforcement learning and imitation learning. We explore using diffusion as a model class for the successor state measure (SSM) of a policy. We find that enforcing the Bellman flow constraints leads to a simple Bellman update on the diffusion step distribution.
View on arXivComments on this paper
