216

Complexity Analysis of Normalizing Constant Estimation: from Jarzynski Equality to Annealed Importance Sampling and beyond

Main:9 Pages
3 Figures
Bibliography:7 Pages
1 Tables
Appendix:29 Pages
Abstract

Given an unnormalized probability density πeV\pi\propto\mathrm{e}^{-V}, estimating its normalizing constant Z=RdeV(x)dxZ=\int_{\mathbb{R}^d}\mathrm{e}^{-V(x)}\mathrm{d}x or free energy F=logZF=-\log Z is a crucial problem in Bayesian statistics, statistical mechanics, and machine learning. It is challenging especially in high dimensions or when π\pi is multimodal. To mitigate the high variance of conventional importance sampling estimators, annealing-based methods such as Jarzynski equality and annealed importance sampling are commonly adopted, yet their quantitative complexity guarantees remain largely unexplored. We take a first step toward a non-asymptotic analysis of annealed importance sampling. In particular, we derive an oracle complexity of O~(dβ2A2ε4)\widetilde{O}\left(\frac{d\beta^2{\mathcal{A}}^2}{\varepsilon^4}\right) for estimating ZZ within ε\varepsilon relative error with high probability, where β\beta is the smoothness of VV and A\mathcal{A} denotes the action of a curve of probability measures interpolating π\pi and a tractable reference distribution. Our analysis, leveraging Girsanov theorem and optimal transport, does not explicitly require isoperimetric assumptions on the target distribution. Finally, to tackle the large action of the widely used geometric interpolation of probability distributions, we propose a new normalizing constant estimation algorithm based on reverse diffusion samplers and establish a framework for analyzing its complexity.

View on arXiv
Comments on this paper