14
9

Accelerating Hamiltonian Monte Carlo via Chebyshev Integration Time

Abstract

Hamiltonian Monte Carlo (HMC) is a popular method in sampling. While there are quite a few works of studying this method on various aspects, an interesting question is how to choose its integration time to achieve acceleration. In this work, we consider accelerating the process of sampling from a distribution π(x)exp(f(x))\pi(x) \propto \exp(-f(x)) via HMC via time-varying integration time. When the potential ff is LL-smooth and mm-strongly convex, i.e.\ for sampling from a log-smooth and strongly log-concave target distribution π\pi, it is known that under a constant integration time, the number of iterations that ideal HMC takes to get an ϵ\epsilon Wasserstein-2 distance to the target π\pi is O(κlog1ϵ)O( \kappa \log \frac{1}{\epsilon} ), where κ:=Lm\kappa := \frac{L}{m} is the condition number. We propose a scheme of time-varying integration time based on the roots of Chebyshev polynomials. We show that in the case of quadratic potential ff, i.e., when the target π\pi is a Gaussian distribution, ideal HMC with this choice of integration time only takes O(κlog1ϵ)O( \sqrt{\kappa} \log \frac{1}{\epsilon} ) number of iterations to reach Wasserstein-2 distance less than ϵ\epsilon; this improvement on the dependence on condition number is akin to acceleration in optimization. The design and analysis of HMC with the proposed integration time is built on the tools of Chebyshev polynomials. Experiments find the advantage of adopting our scheme of time-varying integration time even for sampling from distributions with smooth strongly convex potentials that are not quadratic.

View on arXiv
Comments on this paper