35

Underdamped Langevin MCMC with third order convergence

Main:19 Pages
7 Figures
Bibliography:4 Pages
1 Tables
Appendix:39 Pages
Abstract

In this paper, we propose a new numerical method for the underdamped Langevin diffusion (ULD) and present a non-asymptotic analysis of its sampling error in the 2-Wasserstein distance when the dd-dimensional target distribution p(x)ef(x)p(x)\propto e^{-f(x)} is strongly log-concave and has varying degrees of smoothness. Precisely, under the assumptions that the gradient and Hessian of ff are Lipschitz continuous, our algorithm achieves a 2-Wasserstein error of ε\varepsilon in O(d/ε)\mathcal{O}(\sqrt{d}/\varepsilon) and O(d/ε)\mathcal{O}(\sqrt{d}/\sqrt{\varepsilon}) steps respectively. Therefore, our algorithm has a similar complexity as other popular Langevin MCMC algorithms under matching assumptions. However, if we additionally assume that the third derivative of ff is Lipschitz continuous, then our algorithm achieves a 2-Wasserstein error of ε\varepsilon in O(d/ε13)\mathcal{O}(\sqrt{d}/\varepsilon^{\frac{1}{3}}) steps. To the best of our knowledge, this is the first gradient-only method for ULD with third order convergence. To support our theory, we perform Bayesian logistic regression across a range of real-world datasets, where our algorithm achieves competitive performance compared to an existing underdamped Langevin MCMC algorithm and the popular No U-Turn Sampler (NUTS).

View on arXiv
Comments on this paper