ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1908.11848
  4. Cited By
Dynamic Stale Synchronous Parallel Distributed Training for Deep
  Learning

Dynamic Stale Synchronous Parallel Distributed Training for Deep Learning

IEEE International Conference on Distributed Computing Systems (ICDCS), 2019
16 August 2019
Xing Zhao
Aijun An
Junfeng Liu
Bin Chen
ArXiv (abs)PDFHTML

Papers citing "Dynamic Stale Synchronous Parallel Distributed Training for Deep Learning"

18 / 18 papers shown
Hybrid Dual-Batch and Cyclic Progressive Learning for Efficient Distributed Training
Hybrid Dual-Batch and Cyclic Progressive Learning for Efficient Distributed Training
Kuan-Wei Lu
Ding-Yong Hong
Pangfeng Liu
Jan-Jan Wu
153
0
0
30 Sep 2025
Dynamic Clustering for Personalized Federated Learning on Heterogeneous Edge Devices
Dynamic Clustering for Personalized Federated Learning on Heterogeneous Edge Devices
Heting Liu
Junzhe Huang
Fang He
Guohong Cao
FedML
205
0
0
03 Aug 2025
Acceleration for Deep Reinforcement Learning using Parallel and
  Distributed Computing: A Survey
Acceleration for Deep Reinforcement Learning using Parallel and Distributed Computing: A SurveyACM Computing Surveys (ACM CSUR), 2024
Zhihong Liu
Xin Xu
Peng Qiao
Dongsheng Li
OffRL
336
20
0
08 Nov 2024
Malleus: Straggler-Resilient Hybrid Parallel Training of Large-scale
  Models via Malleable Data and Model Parallelization
Malleus: Straggler-Resilient Hybrid Parallel Training of Large-scale Models via Malleable Data and Model Parallelization
Haoyang Li
Fangcheng Fu
Hao Ge
Sheng Lin
Xuanyu Wang
Jiawen Niu
Yijiao Wang
Hailin Zhang
Xiaonan Nie
Tengjiao Wang
MoMe
419
11
0
17 Oct 2024
An Interdisciplinary Outlook on Large Language Models for Scientific
  Research
An Interdisciplinary Outlook on Large Language Models for Scientific Research
James Boyko
Joseph Cohen
Nathan Fox
Maria Han Veiga
Jennifer I-Hsiu Li
...
Andreas H. Rauch
Kenneth N. Reid
Soumi Tribedi
Anastasia Visheratina
Xin Xie
288
25
0
03 Nov 2023
ABS-SGD: A Delayed Synchronous Stochastic Gradient Descent Algorithm
  with Adaptive Batch Size for Heterogeneous GPU Clusters
ABS-SGD: A Delayed Synchronous Stochastic Gradient Descent Algorithm with Adaptive Batch Size for Heterogeneous GPU Clusters
Xin Zhou
Ling Chen
Houming Wu
192
1
0
29 Aug 2023
OSP: Boosting Distributed Model Training with 2-stage Synchronization
OSP: Boosting Distributed Model Training with 2-stage SynchronizationInternational Conference on Parallel Processing (ICPP), 2023
Zixuan Chen
Lei Shi
Xuandong Liu
Jiahui Li
Sen Liu
Yang Xu
327
5
0
29 Jun 2023
Communication-Efficient Federated Learning for Heterogeneous Edge
  Devices Based on Adaptive Gradient Quantization
Communication-Efficient Federated Learning for Heterogeneous Edge Devices Based on Adaptive Gradient QuantizationIEEE Conference on Computer Communications (INFOCOM), 2022
Heting Liu
Fang He
Guohong Cao
FedMLMQ
205
50
0
16 Dec 2022
A Comprehensive Survey on Distributed Training of Graph Neural Networks
A Comprehensive Survey on Distributed Training of Graph Neural NetworksProceedings of the IEEE (Proc. IEEE), 2022
Haiyang Lin
Yurui Lai
Xiaochun Ye
Xiaochun Ye
Shirui Pan
Wenguang Chen
Yuan Xie
GNN
351
42
0
10 Nov 2022
Fair and Efficient Distributed Edge Learning with Hybrid Multipath TCP
Fair and Efficient Distributed Edge Learning with Hybrid Multipath TCPIEEE/ACM Transactions on Networking (TON), 2022
Mengyue Deng
Jinho Choi
A. Walid
218
10
0
03 Nov 2022
Byzantine Fault Tolerance in Distributed Machine Learning : a Survey
Byzantine Fault Tolerance in Distributed Machine Learning : a Survey
Djamila Bouhata
Hamouma Moumen
Moumen Hamouma
Ahcène Bounceur
AI4CE
383
14
0
05 May 2022
On the Future of Cloud Engineering
On the Future of Cloud Engineering
David Bermbach
A. Chandra
C. Krintz
A. Gokhale
Aleksander Slominski
L. Thamsen
Everton Cavalcante
Tian Guo
Ivona Brandić
R. Wolski
167
28
0
19 Aug 2021
Quantifying and Improving Performance of Distributed Deep Learning with
  Cloud Storage
Quantifying and Improving Performance of Distributed Deep Learning with Cloud Storage
Nicholas Krichevsky
M. S. Louis
Tian Guo
203
9
0
13 Aug 2021
Sync-Switch: Hybrid Parameter Synchronization for Distributed Deep
  Learning
Sync-Switch: Hybrid Parameter Synchronization for Distributed Deep LearningIEEE International Conference on Distributed Computing Systems (ICDCS), 2021
Shijian Li
Oren Mangoubi
Lijie Xu
Tian Guo
280
22
0
16 Apr 2021
BaPipe: Exploration of Balanced Pipeline Parallelism for DNN Training
BaPipe: Exploration of Balanced Pipeline Parallelism for DNN Training
Letian Zhao
Rui Xu
Tianqi Wang
Teng Tian
Xiaotian Wang
Wei Wu
Chio-in Ieong
Xi Jin
MoE
274
8
0
23 Dec 2020
A Fast Edge-Based Synchronizer for Tasks in Real-Time Artificial
  Intelligence Applications
A Fast Edge-Based Synchronizer for Tasks in Real-Time Artificial Intelligence ApplicationsIEEE Internet of Things Journal (IEEE IoT J.), 2020
R. Olaniyan
Muthucumaru Maheswaran
117
3
0
21 Dec 2020
A Quantitative Survey of Communication Optimizations in Distributed Deep
  Learning
A Quantitative Survey of Communication Optimizations in Distributed Deep Learning
Shaoshuai Shi
Zhenheng Tang
Xiaowen Chu
Chengjian Liu
Wei Wang
Bo Li
GNNAI4CE
164
3
0
27 May 2020
Elastic Bulk Synchronous Parallel Model for Distributed Deep Learning
Elastic Bulk Synchronous Parallel Model for Distributed Deep LearningIndustrial Conference on Data Mining (IDM), 2019
Xing Zhao
Manos Papagelis
Aijun An
Bin Chen
Junfeng Liu
Yonggang Hu
260
16
0
06 Jan 2020
1
Page 1 of 1