241

The shifted ODE method for underdamped Langevin MCMC

Abstract

In this paper, we consider the underdamped Langevin diffusion (ULD) and propose a numerical approximation using its associated ordinary differential equation (ODE). When used as a Markov Chain Monte Carlo (MCMC) algorithm, we show that the ODE approximation achieves a 22-Wasserstein error of ε\varepsilon in O(d13/ε23)\mathcal{O}\big(d^{\frac{1}{3}}/\varepsilon^{\frac{2}{3}}\big) steps under the standard smoothness and strong convexity assumptions on the target distribution. This matches the complexity of the randomized midpoint method proposed by Shen and Lee [NeurIPS 2019] which was shown to be order optimal by Cao, Lu and Wang. However, the main feature of the proposed numerical method is that it can utilize additional smoothness of the target log-density ff. More concretely, we show that the ODE approximation achieves a 22-Wasserstein error of ε\varepsilon in O(d25/ε25)\mathcal{O}\big(d^{\frac{2}{5}}/\varepsilon^{\frac{2}{5}}\big) and O(d/ε13)\mathcal{O}\big(\sqrt{d}/\varepsilon^{\frac{1}{3}}\big) steps when Lipschitz continuity is assumed for the Hessian and third derivative of ff. By discretizing this ODE using a fourth order splitting method, we obtain a practical MCMC method that requires just three additional gradient evaluations in each step. In our experiment, where the target comes from a logistic regression, this method shows faster convergence compared to other unadjusted Langevin MCMC algorithms.

View on arXiv
Comments on this paper