ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2301.05872
17
9

CEDAS: A Compressed Decentralized Stochastic Gradient Method with Improved Convergence

14 January 2023
Kun-Yen Huang
Shin-Yi Pu
ArXivPDFHTML
Abstract

In this paper, we consider solving the distributed optimization problem over a multi-agent network under the communication restricted setting. We study a compressed decentralized stochastic gradient method, termed ``compressed exact diffusion with adaptive stepsizes (CEDAS)", and show the method asymptotically achieves comparable convergence rate as centralized { stochastic gradient descent (SGD)} for both smooth strongly convex objective functions and smooth nonconvex objective functions under unbiased compression operators. In particular, to our knowledge, CEDAS enjoys so far the shortest transient time (with respect to the graph specifics) for achieving the convergence rate of centralized SGD, which behaves as O(nC3/(1−λ2)2)\mathcal{O}(n{C^3}/(1-\lambda_2)^{2})O(nC3/(1−λ2​)2) under smooth strongly convex objective functions, and O(n3C6/(1−λ2)4)\mathcal{O}(n^3{C^6}/(1-\lambda_2)^4)O(n3C6/(1−λ2​)4) under smooth nonconvex objective functions, where (1−λ2)(1-\lambda_2)(1−λ2​) denotes the spectral gap of the mixing matrix, and C>0C>0C>0 is the compression-related parameter. Numerical experiments further demonstrate the effectiveness of the proposed algorithm.

View on arXiv
Comments on this paper