ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2105.09080
  4. Cited By
Accelerating Gossip SGD with Periodic Global Averaging

Accelerating Gossip SGD with Periodic Global Averaging

International Conference on Machine Learning (ICML), 2021
19 May 2021
Yiming Chen
Kun Yuan
Yingya Zhang
Pan Pan
Yinghui Xu
W. Yin
ArXiv (abs)PDFHTML

Papers citing "Accelerating Gossip SGD with Periodic Global Averaging"

28 / 28 papers shown
Nesterov-Accelerated Robust Federated Learning Over Byzantine Adversaries
Nesterov-Accelerated Robust Federated Learning Over Byzantine Adversaries
Lihan Xu
Yanjie Dong
Gang Wang
Runhao Zeng
Xiaoyi Fan
Xiping Hu
FedMLAAML
275
1
0
04 Nov 2025
On the Surprising Effectiveness of a Single Global Merging in Decentralized Learning
On the Surprising Effectiveness of a Single Global Merging in Decentralized Learning
Tongtian Zhu
Tianyu Zhang
Mingze Wang
Zhanpeng Zhou
Can Wang
FedML
384
0
0
09 Jul 2025
Distribution-Aware Mobility-Assisted Decentralized Federated Learning
Distribution-Aware Mobility-Assisted Decentralized Federated Learning
Md. Farhamdur Reza
Reza Jahani
Richeng Jin
H. Dai
229
1
0
24 May 2025
Decentralized Low-Rank Fine-Tuning of Large Language Models
Decentralized Low-Rank Fine-Tuning of Large Language Models
Sajjad Ghiasvand
Mahnoosh Alizadeh
Ramtin Pedarsani
ALM
712
12
0
26 Jan 2025
EDiT: A Local-SGD-Based Efficient Distributed Training Method for Large Language Models
EDiT: A Local-SGD-Based Efficient Distributed Training Method for Large Language ModelsInternational Conference on Learning Representations (ICLR), 2024
Jialiang Cheng
Ning Gao
Yun Yue
Zhiling Ye
Jiadi Jiang
Jian Sha
OffRL
541
1
0
10 Dec 2024
Adjacent Leader Decentralized Stochastic Gradient Descent
Adjacent Leader Decentralized Stochastic Gradient DescentEuropean Conference on Artificial Intelligence (ECAI), 2024
Haoze He
Jing Wang
A. Choromańska
235
0
0
18 May 2024
Vanishing Variance Problem in Fully Decentralized Neural-Network Systems
Vanishing Variance Problem in Fully Decentralized Neural-Network Systems
Yongding Tian
Zaid Al-Ars
Maksim Kitsak
P. Hofstee
FedML
244
2
0
06 Apr 2024
The Effectiveness of Local Updates for Decentralized Learning under Data
  Heterogeneity
The Effectiveness of Local Updates for Decentralized Learning under Data HeterogeneityIEEE Transactions on Signal Processing (IEEE TSP), 2024
Tongle Wu
Ying Sun
229
6
0
23 Mar 2024
Decentralized Bilevel Optimization: A Perspective from Transient Iteration Complexity
Decentralized Bilevel Optimization: A Perspective from Transient Iteration Complexity
Boao Kong
Shuchen Zhu
Songtao Lu
Xinmeng Huang
Kun Yuan
476
0
0
05 Feb 2024
Communication-Efficient Federated Optimization over Semi-Decentralized Networks
Communication-Efficient Federated Optimization over Semi-Decentralized NetworksIEEE Transactions on Signal and Information Processing over Networks (TSIPN), 2023
He Wang
Yuejie Chi
FedML
577
6
0
30 Nov 2023
Over-the-Air Federated Learning and Optimization
Over-the-Air Federated Learning and Optimization
Jingyang Zhu
Yuanming Shi
Yong Zhou
Chunxiao Jiang
Wei Chen
Khaled B. Letaief
FedML
547
29
0
16 Oct 2023
Epidemic Learning: Boosting Decentralized Learning with Randomized
  Communication
Epidemic Learning: Boosting Decentralized Learning with Randomized CommunicationNeural Information Processing Systems (NeurIPS), 2023
M. Vos
Sadegh Farhadkhani
R. Guerraoui
Anne-Marie Kermarrec
Rafael Pires
Rishi Sharma
366
29
0
03 Oct 2023
Decentralized Local Updates with Dual-Slow Estimation and Momentum-based
  Variance-Reduction for Non-Convex Optimization
Decentralized Local Updates with Dual-Slow Estimation and Momentum-based Variance-Reduction for Non-Convex OptimizationEuropean Conference on Artificial Intelligence (ECAI), 2023
Kangyang Luo
Kunkun Zhang
Sheng Zhang
Xiang Li
Ming Gao
156
2
0
17 Jul 2023
FedDec: Peer-to-peer Aided Federated Learning
FedDec: Peer-to-peer Aided Federated LearningInternational Workshop on Signal Processing Advances in Wireless Communications (SPAWC), 2023
Marina Costantini
Giovanni Neglia
T. Spyropoulos
FedML
145
2
0
11 Jun 2023
Beyond Exponential Graph: Communication-Efficient Topologies for
  Decentralized Learning via Finite-time Convergence
Beyond Exponential Graph: Communication-Efficient Topologies for Decentralized Learning via Finite-time ConvergenceNeural Information Processing Systems (NeurIPS), 2023
Yuki Takezawa
Ryoma Sato
Han Bao
Kenta Niwa
M. Yamada
349
18
0
19 May 2023
On the Limit Performance of Floating Gossip
On the Limit Performance of Floating GossipIEEE Conference on Computer Communications (IEEE INFOCOM), 2023
Gianluca Rizzo
Noelia Pérez Palma
M. Marsan
Vincenzo Mancuso
FedML
171
0
0
16 Feb 2023
Decentralized Gradient Tracking with Local Steps
Decentralized Gradient Tracking with Local StepsOptimization Methods and Software (OMS), 2023
Yue Liu
Tao Lin
Anastasia Koloskova
Sebastian U. Stich
291
61
0
03 Jan 2023
Revisiting Optimal Convergence Rate for Smooth and Non-convex Stochastic
  Decentralized Optimization
Revisiting Optimal Convergence Rate for Smooth and Non-convex Stochastic Decentralized OptimizationNeural Information Processing Systems (NeurIPS), 2022
Kun Yuan
Xinmeng Huang
Yiming Chen
Xiaohan Zhang
Yingya Zhang
Pan Pan
286
33
0
14 Oct 2022
Tackling Data Heterogeneity: A New Unified Framework for Decentralized
  SGD with Sample-induced Topology
Tackling Data Heterogeneity: A New Unified Framework for Decentralized SGD with Sample-induced TopologyInternational Conference on Machine Learning (ICML), 2022
Yan Huang
Ying Sun
Zehan Zhu
Changzhi Yan
Jinming Xu
FedML
273
18
0
08 Jul 2022
Lower Bounds and Nearly Optimal Algorithms in Distributed Learning with
  Communication Compression
Lower Bounds and Nearly Optimal Algorithms in Distributed Learning with Communication CompressionNeural Information Processing Systems (NeurIPS), 2022
Xinmeng Huang
Yiming Chen
W. Yin
Kun Yuan
420
40
0
08 Jun 2022
Sign Bit is Enough: A Learning Synchronization Framework for Multi-hop
  All-reduce with Ultimate Compression
Sign Bit is Enough: A Learning Synchronization Framework for Multi-hop All-reduce with Ultimate CompressionDesign Automation Conference (DAC), 2022
Feijie Wu
Shiqi He
Song Guo
Zhihao Qu
Yining Qi
W. Zhuang
Jie Zhang
220
10
0
14 Apr 2022
Decentralized Optimization Over the Stiefel Manifold by an Approximate
  Augmented Lagrangian Function
Decentralized Optimization Over the Stiefel Manifold by an Approximate Augmented Lagrangian FunctionIEEE Transactions on Signal Processing (IEEE TSP), 2021
Lei Wang
Xin Liu
286
34
0
30 Dec 2021
BlueFog: Make Decentralized Algorithms Practical for Optimization and
  Deep Learning
BlueFog: Make Decentralized Algorithms Practical for Optimization and Deep Learning
Bicheng Ying
Kun Yuan
Hanbin Hu
Yiming Chen
W. Yin
FedML
299
31
0
08 Nov 2021
Exponential Graph is Provably Efficient for Decentralized Deep Training
Exponential Graph is Provably Efficient for Decentralized Deep Training
Bicheng Ying
Kun Yuan
Yiming Chen
Hanbin Hu
Pan Pan
W. Yin
FedML
262
109
0
26 Oct 2021
A Unified and Refined Convergence Analysis for Non-Convex Decentralized
  Learning
A Unified and Refined Convergence Analysis for Non-Convex Decentralized Learning
Sulaiman A. Alghunaim
Kun Yuan
258
82
0
19 Oct 2021
Removing Data Heterogeneity Influence Enhances Network Topology
  Dependence of Decentralized SGD
Removing Data Heterogeneity Influence Enhances Network Topology Dependence of Decentralized SGDJournal of machine learning research (JMLR), 2021
Kun Yuan
Sulaiman A. Alghunaim
Xinmeng Huang
264
41
0
17 May 2021
DecentLaM: Decentralized Momentum SGD for Large-batch Deep Training
DecentLaM: Decentralized Momentum SGD for Large-batch Deep TrainingIEEE International Conference on Computer Vision (ICCV), 2021
Kun Yuan
Yiming Chen
Xinmeng Huang
Yingya Zhang
Pan Pan
Yinghui Xu
W. Yin
MoE
259
71
0
24 Apr 2021
Crossover-SGD: A gossip-based communication in distributed deep learning
  for alleviating large mini-batch problem and enhancing scalability
Crossover-SGD: A gossip-based communication in distributed deep learning for alleviating large mini-batch problem and enhancing scalabilityConcurrency and Computation (CCPE), 2020
Sangho Yeo
Minho Bae
Minjoong Jeong
Oh-Kyoung Kwon
Sangyoon Oh
256
4
0
30 Dec 2020
1
Page 1 of 1