In this work, we establish the linear convergence estimate for the gradient descent involving the delay when the cost function is -strongly convex and -smooth. This result improves upon the well-known estimates in Arjevani et al. \cite{ASS} and Stich-Karmireddy \cite{SK} in the sense that it is non-ergodic and is still established in spite of weaker constraint of cost function. Also, the range of learning rate can be extended from to for and for , where is the Lipschitz continuity constant of the gradient of cost function. In a further research, we show the linear convergence of cost function under the Polyak-{\L}ojasiewicz\,(PL) condition, for which the available choice of learning rate is further improved as for the large delay . Finally, some numerical experiments are provided in order to confirm the reliability of the analyzed results.
View on arXiv