348

Black-box αα-divergence Minimization

Abstract

We present \emph{black-box alpha} (BB-α\alpha), an approximate inference method based on the minimization of α\alpha-divergences between probability distributions. BB-α\alpha scales to large datasets since it can be implemented using stochastic gradient descent. BB-α\alpha can be applied to complex probabilistic models with little effort since it only requires as input the likelihood function and its gradients. These gradients can be easily obtained using automatic differentiation. By tuning divergence parameter α\alpha, the method is able to interpolate between variational Bayes and an expectation propagation-like algorithm. Experiments on probit regression, neural network regression and classification problems illustrate the accuracy of the posterior approximations obtained with BB-α\alpha.

View on arXiv
Comments on this paper