167

An Optimal Stochastic Algorithm for Decentralized Nonconvex Finite-sum Optimization

Main:14 Pages
4 Figures
Bibliography:3 Pages
1 Tables
Abstract

This paper studies the synchronized decentralized nonconvex optimization problem of the form minxRdf(x)1mi=1mfi(x)\min_{x\in{\mathbb R}^d} f(x)\triangleq \frac{1}{m}\sum_{i=1}^m f_i(x), where fi(x)1nj=1nfi,j(x)f_i(x)\triangleq \frac{1}{n}\sum_{j=1}^n f_{i,j}(x) is the local function on ii-th agent of the connected network. We propose a novel stochastic algorithm called DEcentralized probAbilistic Recursive gradiEnt deScenT (DEAREST), which integrates the techniques of variance reduction, gradient tracking and multi-consensus. We construct a Lyapunov function that simultaneously characterizes the function value, the gradient estimation error and the consensus error for the convergence analysis. Based on this measure, we provide a concise proof to show DEAREST requires at most O(mn+mnLε2){\mathcal O}(mn+\sqrt{mn}L\varepsilon^{-2}) incremental first-order oracle (IFO) calls and O(Lε2/1λ2(W)){\mathcal O}(L\varepsilon^{-2}/\sqrt{1-\lambda_2(W)}\,) communication rounds to find an ε\varepsilon-stationary point in expectation, where LL is the smoothness parameter and λ2(W)\lambda_2(W) is the second-largest eigenvalues of the gossip matrix WW. We can verify both of the IFO complexity and communication complexity match the lower bounds. To the best of our knowledge, DEAREST is the first optimal algorithm for decentralized nonconvex finite-sum optimization.

View on arXiv
Comments on this paper