Sequential Monte Carlo approximations of Wasserstein--Fisher--Rao gradient flows

We consider the problem of sampling from a probability distribution . It is well known that this can be written as an optimisation problem over the space of probability distribution in which we aim to minimise the Kullback--Leibler divergence from . We consider several partial differential equations (PDEs) whose solution is a minimiser of the Kullback--Leibler divergence from and connect them to well-known Monte Carlo algorithms. We focus in particular on PDEs obtained by considering the Wasserstein--Fisher--Rao geometry over the space of probabilities and show that these lead to a natural implementation using importance sampling and sequential Monte Carlo. We propose a novel algorithm to approximate the Wasserstein--Fisher--Rao flow of the Kullback--Leibler divergence which empirically outperforms the current state-of-the-art.We study tempered versions of these PDEs obtained by replacing the target distribution with a geometric mixture of initial and target distribution and show that these do not lead to a convergence speed up.
View on arXiv@article{crucinio2025_2506.05905, title={ Sequential Monte Carlo approximations of Wasserstein--Fisher--Rao gradient flows }, author={ Francesca R. Crucinio and Sahani Pathiraja }, journal={arXiv preprint arXiv:2506.05905}, year={ 2025 } }