ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2110.13144
18
1

Faster Perturbed Stochastic Gradient Methods for Finding Local Minima

25 October 2021
Zixiang Chen
Dongruo Zhou
Quanquan Gu
ArXivPDFHTML
Abstract

Escaping from saddle points and finding local minimum is a central problem in nonconvex optimization. Perturbed gradient methods are perhaps the simplest approach for this problem. However, to find (ϵ,ϵ)(\epsilon, \sqrt{\epsilon})(ϵ,ϵ​)-approximate local minima, the existing best stochastic gradient complexity for this type of algorithms is O~(ϵ−3.5)\tilde O(\epsilon^{-3.5})O~(ϵ−3.5), which is not optimal. In this paper, we propose LENA (Last stEp shriNkAge), a faster perturbed stochastic gradient framework for finding local minima. We show that LENA with stochastic gradient estimators such as SARAH/SPIDER and STORM can find (ϵ,ϵH)(\epsilon, \epsilon_{H})(ϵ,ϵH​)-approximate local minima within O~(ϵ−3+ϵH−6)\tilde O(\epsilon^{-3} + \epsilon_{H}^{-6})O~(ϵ−3+ϵH−6​) stochastic gradient evaluations (or O~(ϵ−3)\tilde O(\epsilon^{-3})O~(ϵ−3) when ϵH=ϵ\epsilon_H = \sqrt{\epsilon}ϵH​=ϵ​). The core idea of our framework is a step-size shrinkage scheme to control the average movement of the iterates, which leads to faster convergence to the local minima.

View on arXiv
Comments on this paper