64
2

Sampling from Log-Concave Distributions over Polytopes via a Soft-Threshold Dikin Walk

Abstract

Given a Lipschitz or smooth convex function f:KR\, f:K \to \mathbb{R} for a bounded polytope KRdK \subseteq \mathbb{R}^d defined by mm inequalities, we consider the problem of sampling from the log-concave distribution π(θ)ef(θ)\pi(\theta) \propto e^{-f(\theta)} constrained to KK. Interest in this problem derives from its applications to Bayesian inference and differentially private learning. Our main result is a generalization of the Dikin walk Markov chain to this setting that requires at most O((md+dL2R2)×mdω1)log(wδ))O((md + d L^2 R^2) \times md^{\omega-1}) \log(\frac{w}{\delta})) arithmetic operations to sample from π\pi within error δ>0\delta>0 in the total variation distance from a ww-warm start. Here LL is the Lipschitz-constant of ff, KK is contained in a ball of radius RR and contains a ball of smaller radius rr, and ω\omega is the matrix-multiplication constant. Our algorithm improves on the running time of prior works for a range of parameter settings important for the aforementioned learning applications. Technically, we depart from previous Dikin walks by adding a "soft-threshold" regularizer derived from the Lipschitz or smoothness properties of ff to the log-barrier function for KK that allows our version of the Dikin walk to propose updates that have a high Metropolis acceptance ratio for ff, while at the same time remaining inside the polytope KK.

View on arXiv
Comments on this paper