ResearchTrend.AI
  • Papers
  • Communities
  • Organizations
  • Events
  • Blog
  • Pricing
  • Feedback
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1707.06618
142
212
v1v2v3 (latest)

Global Convergence of Langevin Dynamics Based Algorithms for Nonconvex Optimization

20 July 2017
Pan Xu
Jinghui Chen
Difan Zou
Quanquan Gu
ArXiv (abs)PDFHTML
Abstract

We present a unified framework to analyze the global convergence of Langevin dynamics based algorithms for nonconvex finite-sum optimization with nnn component functions. At the core of our analysis is a direct analysis of the ergodicity of the numerical approximations to Langevin dynamics, which leads to faster convergence rates. Specifically, we show that gradient Langevin dynamics (GLD) and stochastic gradient Langevin dynamics (SGLD) converge to the almost minimizer within O~(nd/(λϵ))\tilde O\big(nd/(\lambda\epsilon) \big)O~(nd/(λϵ)) and O~(d7/(λ5ϵ5))\tilde O\big(d^7/(\lambda^5\epsilon^5) \big)O~(d7/(λ5ϵ5)) stochastic gradient evaluations respectively, where ddd is the problem dimension, and λ\lambdaλ is the spectral gap of the Markov chain generated by GLD. Both of the results improve upon the best known gradient complexity results. Furthermore, for the first time we prove the global convergence guarantee for variance reduced stochastic gradient Langevin dynamics (VR-SGLD) to the almost minimizer after O~(nd5/(λ4ϵ5/2))\tilde O\big(\sqrt{n}d^5/(\lambda^4\epsilon^{5/2})\big)O~(n​d5/(λ4ϵ5/2)) stochastic gradient evaluations, which outperforms the gradient complexities of GLD and SGLD in a wide regime. Our theoretical analyses shed some light on using Langevin dynamics based algorithms for nonconvex optimization with provable guarantees.

View on arXiv
Comments on this paper