ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1902.08179
31
3
v1v2v3v4 (latest)

Online Sampling from Log-Concave Distributions

21 February 2019
Holden Lee
Oren Mangoubi
Nisheeth K. Vishnoi
ArXiv (abs)PDFHTML
Abstract

Given a sequence of convex functions f0,f1,…,fTf_0, f_1, \ldots, f_Tf0​,f1​,…,fT​, we study the problem of sampling from the Gibbs distribution πt∝e−∑k=0tfk\pi_t \propto e^{-\sum_{k=0}^tf_k}πt​∝e−∑k=0t​fk​ for each epoch ttt in an online manner. Interest in this problem derives from applications in machine learning, Bayesian statistics, and optimization where, rather than obtaining all the observations at once, one constantly acquires new data, and must continuously update the distribution. Our main result is an algorithm that generates roughly independent samples from πt\pi_tπt​ for every epoch ttt and, under mild assumptions, makes polylog(T)\mathrm{polylog}(T)polylog(T) gradient evaluations per epoch. All previous results imply a bound on the number of gradient or function evaluations which is at least linear in TTT. Motivated by real-world applications, we assume that functions are smooth, their associated distributions have a bounded second moment, and their minimizer drifts in a bounded manner, but do not assume they are strongly convex. In particular, our assumptions hold for online Bayesian logistic regression, when the data satisfy natural regularity properties, giving a sampling algorithm with updates that are poly-logarithmic in TTT. In simulations, our algorithm achieves accuracy comparable to an algorithm specialized to logistic regression. Key to our algorithm is a novel stochastic gradient Langevin dynamics Markov chain with a carefully designed variance reduction step and constant batch size. Technically, lack of strong convexity is a significant barrier to analysis and, here, our main contribution is a martingale exit time argument that shows our Markov chain remains in a ball of radius roughly poly-logarithmic in TTT for enough time to reach within ε\varepsilonε of πt\pi_tπt​.

View on arXiv
Comments on this paper