ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2308.11984
46
0
v1v2 (latest)

Non-ergodic linear convergence property of the delayed gradient descent under the strongly convexity and the Polyak-Łojasiewicz condition

23 August 2023
Hyunggwon Choi
Woocheol Choi
Jinmyoung Seok
ArXiv (abs)PDFHTML
Abstract

In this work, we establish the linear convergence estimate for the gradient descent involving the delay τ∈N\tau\in\mathbb{N}τ∈N when the cost function is μ\muμ-strongly convex and LLL-smooth. This result improves upon the well-known estimates in Arjevani et al. \cite{ASS} and Stich-Karmireddy \cite{SK} in the sense that it is non-ergodic and is still established in spite of weaker constraint of cost function. Also, the range of learning rate η\etaη can be extended from η≤1/(10Lτ)\eta\leq 1/(10L\tau)η≤1/(10Lτ) to η≤1/(4Lτ)\eta\leq 1/(4L\tau)η≤1/(4Lτ) for τ=1\tau =1τ=1 and η≤3/(10Lτ)\eta\leq 3/(10L\tau)η≤3/(10Lτ) for τ≥2\tau \geq 2τ≥2, where L>0L >0L>0 is the Lipschitz continuity constant of the gradient of cost function. In a further research, we show the linear convergence of cost function under the Polyak-{\L}ojasiewicz\,(PL) condition, for which the available choice of learning rate is further improved as η≤9/(10Lτ)\eta\leq 9/(10L\tau)η≤9/(10Lτ) for the large delay τ\tauτ. Finally, some numerical experiments are provided in order to confirm the reliability of the analyzed results.

View on arXiv
Comments on this paper