108

Marginal Flow: a flexible and efficient framework for density estimation

Main:10 Pages
18 Figures
Bibliography:3 Pages
2 Tables
Appendix:6 Pages
Abstract

Current density modeling approaches suffer from at least one of the following shortcomings: expensive training, slow inference, approximate likelihood, mode collapse or architectural constraints like bijective mappings. We propose a simple yet powerful framework that overcomes these limitations altogether. We define our model qθ(x)q_\theta(x) through a parametric distribution q(xw)q(x|w) with latent parameters ww. Instead of directly optimizing the latent variables ww, our idea is to marginalize them out by sampling ww from a learnable distribution qθ(w)q_\theta(w), hence the name Marginal Flow. In order to evaluate the learned density qθ(x)q_\theta(x) or to sample from it, we only need to draw samples from qθ(w)q_\theta(w), which makes both operations efficient. The proposed model allows for exact density evaluation and is orders of magnitude faster than competing models both at training and inference. Furthermore, Marginal Flow is a flexible framework: it does not impose any restrictions on the neural network architecture, it enables learning distributions on lower-dimensional manifolds (either known or to be learned), it can be trained efficiently with any objective (e.g. forward and reverse KL divergence), and it easily handles multi-modal targets. We evaluate Marginal Flow extensively on various tasks including synthetic datasets, simulation-based inference, distributions on positive definite matrices and manifold learning in latent spaces of images.

View on arXiv
Comments on this paper