Convergence of Decentralized Stochastic Subgradient-based Methods for Nonsmooth Nonconvex functions

In this paper, we focus on the decentralized stochastic subgradient-based methods in minimizing nonsmooth nonconvex functions without Clarke regularity, especially in the decentralized training of nonsmooth neural networks. We propose a general framework that unifies various decentralized subgradient-based methods, such as decentralized stochastic subgradient descent (DSGD), DSGD with gradient-tracking technique (DSGD-T), and DSGD with momentum (DSGD-M). To establish the convergence properties of our proposed framework, we relate the discrete iterates to the trajectories of a continuous-time differential inclusion, which is assumed to have a coercive Lyapunov function with a stable set . We prove the asymptotic convergence of the iterates to the stable set with sufficiently small and diminishing step-sizes. These results provide first convergence guarantees for some well-recognized of decentralized stochastic subgradient-based methods without Clarke regularity of the objective function. Preliminary numerical experiments demonstrate that our proposed framework yields highly efficient decentralized stochastic subgradient-based methods with convergence guarantees in the training of nonsmooth neural networks.
View on arXiv@article{zhang2025_2403.11565, title={ Convergence of Decentralized Stochastic Subgradient-based Methods for Nonsmooth Nonconvex functions }, author={ Siyuan Zhang and Nachuan Xiao and Xin Liu }, journal={arXiv preprint arXiv:2403.11565}, year={ 2025 } }