8
32

f-Divergence Variational Inference

Abstract

This paper introduces the ff-divergence variational inference (ff-VI) that generalizes variational inference to all ff-divergences. Initiated from minimizing a crafty surrogate ff-divergence that shares the statistical consistency with the ff-divergence, the ff-VI framework not only unifies a number of existing VI methods, e.g. Kullback-Leibler VI, R\'{e}nyi's α\alpha-VI, and χ\chi-VI, but offers a standardized toolkit for VI subject to arbitrary divergences from ff-divergence family. A general ff-variational bound is derived and provides a sandwich estimate of marginal likelihood (or evidence). The development of the ff-VI unfolds with a stochastic optimization scheme that utilizes the reparameterization trick, importance weighting and Monte Carlo approximation; a mean-field approximation scheme that generalizes the well-known coordinate ascent variational inference (CAVI) is also proposed for ff-VI. Empirical examples, including variational autoencoders and Bayesian neural networks, are provided to demonstrate the effectiveness and the wide applicability of ff-VI.

View on arXiv
Comments on this paper