Rapid Convergence of the Unadjusted Langevin Algorithm: Isoperimetry
Suffices
Neural Information Processing Systems (NeurIPS), 2019
Abstract
We study the Unadjusted Langevin Algorithm (ULA) for sampling from a probability distribution on . We prove a convergence guarantee in Kullback-Leibler (KL) divergence assuming satisfies a log-Sobolev inequality and the Hessian of is bounded. Notably, we do not assume convexity or bounds on higher derivatives. We also prove convergence guarantees in R\'enyi divergence of order assuming the limit of ULA satisfies either the log-Sobolev or Poincar\'e inequality.
View on arXivComments on this paper
