215

Near-Optimal Approximations for Bayesian Inference in Function Space

Main:54 Pages
6 Figures
Bibliography:5 Pages
4 Tables
Abstract

We propose a scalable inference algorithm for Bayes posteriors defined on a reproducing kernel Hilbert space (RKHS). Given a likelihood function and a Gaussian random element representing the prior, the corresponding Bayes posterior measure ΠB\Pi_{\text{B}} can be obtained as the stationary distribution of an RKHS-valued Langevin diffusion. We approximate the infinite-dimensional Langevin diffusion via a projection onto the first MM components of the Kosambi-Karhunen-Loève expansion. Exploiting the thus obtained approximate posterior for these MM components, we perform inference for ΠB\Pi_{\text{B}} by relying on the law of total probability and a sufficiency assumption. The resulting method scales as O(M3+JM2)O(M^3+JM^2), where JJ is the number of samples produced from the posterior measure ΠB\Pi_{\text{B}}. Interestingly, the algorithm recovers the posterior arising from the sparse variational Gaussian process (SVGP) (see Titsias, 2009) as a special case, owed to the fact that the sufficiency assumption underlies both methods. However, whereas the SVGP is parametrically constrained to be a Gaussian process, our method is based on a non-parametric variational family P(RM)\mathcal{P}(\mathbb{R}^M) consisting of all probability measures on RM\mathbb{R}^M. As a result, our method is provably close to the optimal MM-dimensional variational approximation of the Bayes posterior ΠB\Pi_{\text{B}} in P(RM)\mathcal{P}(\mathbb{R}^M) for convex and Lipschitz continuous negative log likelihoods, and coincides with SVGP for the special case of a Gaussian error likelihood.

View on arXiv
Comments on this paper