50
1

Convergence of Decentralized Stochastic Subgradient-based Methods for Nonsmooth Nonconvex functions

Abstract

In this paper, we focus on the decentralized stochastic subgradient-based methods in minimizing nonsmooth nonconvex functions without Clarke regularity, especially in the decentralized training of nonsmooth neural networks. We propose a general framework that unifies various decentralized subgradient-based methods, such as decentralized stochastic subgradient descent (DSGD), DSGD with gradient-tracking technique (DSGD-T), and DSGD with momentum (DSGD-M). To establish the convergence properties of our proposed framework, we relate the discrete iterates to the trajectories of a continuous-time differential inclusion, which is assumed to have a coercive Lyapunov function with a stable set A\mathcal{A}. We prove the asymptotic convergence of the iterates to the stable set A\mathcal{A} with sufficiently small and diminishing step-sizes. These results provide first convergence guarantees for some well-recognized of decentralized stochastic subgradient-based methods without Clarke regularity of the objective function. Preliminary numerical experiments demonstrate that our proposed framework yields highly efficient decentralized stochastic subgradient-based methods with convergence guarantees in the training of nonsmooth neural networks.

View on arXiv
@article{zhang2025_2403.11565,
  title={ Convergence of Decentralized Stochastic Subgradient-based Methods for Nonsmooth Nonconvex functions },
  author={ Siyuan Zhang and Nachuan Xiao and Xin Liu },
  journal={arXiv preprint arXiv:2403.11565},
  year={ 2025 }
}
Comments on this paper