345
v1v2 (latest)

CEDAS: A Compressed Decentralized Stochastic Gradient Method with Improved Convergence

IEEE Transactions on Automatic Control (TAC), 2023
Abstract

In this paper, we consider solving the distributed optimization problem over a multi-agent network under the communication restricted setting. We study a compressed decentralized stochastic gradient method, termed ``compressed exact diffusion with adaptive stepsizes (CEDAS)", and show the method asymptotically achieves comparable convergence rate as centralized { stochastic gradient descent (SGD)} for both smooth strongly convex objective functions and smooth nonconvex objective functions under unbiased compression operators. In particular, to our knowledge, CEDAS enjoys so far the shortest transient time (with respect to the graph specifics) for achieving the convergence rate of centralized SGD, which behaves as O(nC3/(1λ2)2)\mathcal{O}(n{C^3}/(1-\lambda_2)^{2}) under smooth strongly convex objective functions, and O(n3C6/(1λ2)4)\mathcal{O}(n^3{C^6}/(1-\lambda_2)^4) under smooth nonconvex objective functions, where (1λ2)(1-\lambda_2) denotes the spectral gap of the mixing matrix, and C>0C>0 is the compression-related parameter. Numerical experiments further demonstrate the effectiveness of the proposed algorithm.

View on arXiv
Comments on this paper