Penalized Overdamped and Underdamped Langevin Monte Carlo Algorithms for
Constrained Sampling
We consider the constrained sampling problem where the goal is to sample from a target distribution when is constrained to lie on a convex body . Motivated by penalty methods from continuous optimization, we propose penalized Langevin Dynamics (PLD) and penalized underdamped Langevin Monte Carlo (PULMC) methods that convert the constrained sampling problem into an unconstrained sampling problem by introducing a penalty function for constraint violations. When is smooth and gradients are available, we get iteration complexity for PLD to sample the target up to an -error where the error is measured in the TV distance and hides logarithmic factors. For PULMC, we improve the result to when the Hessian of is Lipschitz and the boundary of is sufficiently smooth. To our knowledge, these are the first convergence results for underdamped Langevin Monte Carlo methods in the constrained sampling that handle non-convex and provide guarantees with the best dimension dependency among existing methods with deterministic gradient. If unbiased stochastic estimates of the gradient of are available, we propose PSGLD and PSGULMC methods that can handle stochastic gradients and are scaleable to large datasets without requiring Metropolis-Hasting correction steps. For PSGLD and PSGULMC, when is strongly convex and smooth, we obtain and iteration complexity in W2 distance. When is smooth and can be non-convex, we provide finite-time performance bounds and iteration complexity results. Finally, we illustrate the performance on Bayesian LASSO regression and Bayesian constrained deep learning problems.
View on arXiv