DoubleSqueeze: Parallel Stochastic Gradient Descent with Double-Pass
Error-Compensated CompressionInternational Conference on Machine Learning (ICML), 2019 |
Priority-based Parameter Propagation for Distributed DNN TrainingUSENIX workshop on Tackling computer systems problems with machine learning techniques (SysML), 2019 |
Pipe-SGD: A Decentralized Pipelined SGD Framework for Distributed Deep
Net TrainingNeural Information Processing Systems (NeurIPS), 2018 |
Demystifying Parallel and Distributed Deep Learning: An In-Depth
Concurrency AnalysisACM Computing Surveys (CSUR), 2018 |