16
0

Delayed Stochastic Algorithms for Distributed Weakly Convex Optimization

Abstract

This paper studies delayed stochastic algorithms for weakly convex optimization in a distributed network with workers connected to a master node. Recently, Xu et al. 2022 showed that an inertial stochastic subgradient method converges at a rate of O(τmax/K)\mathcal{O}(\tau_{\text{max}}/\sqrt{K}) which depends on the maximum information delay τmax\tau_{\text{max}}. In this work, we show that the delayed stochastic subgradient method (DSGD\texttt{DSGD}) obtains a tighter convergence rate which depends on the expected delay τˉ\bar{\tau}. Furthermore, for an important class of composition weakly convex problems, we develop a new delayed stochastic prox-linear (DSPL\texttt{DSPL}) method in which the delays only affect the high-order term in the complexity rate and hence, are negligible after a certain number of DSPL\texttt{DSPL} iterations. In addition, we demonstrate the robustness of our proposed algorithms against arbitrary delays. By incorporating a simple safeguarding step in both methods, we achieve convergence rates that depend solely on the number of workers, eliminating the effect of the delay. Our numerical experiments further confirm the empirical superiority of our proposed methods.

View on arXiv
Comments on this paper