Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2302.02390
Cited By
Quantized Distributed Training of Large Models with Convergence Guarantees
5 February 2023
I. Markov
Adrian Vladu
Qi Guo
Dan Alistarh
MQ
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Quantized Distributed Training of Large Models with Convergence Guarantees"
3 / 3 papers shown
Title
ZeRO-Offload: Democratizing Billion-Scale Model Training
Jie Ren
Samyam Rajbhandari
Reza Yazdani Aminabadi
Olatunji Ruwase
Shuangyang Yang
Minjia Zhang
Dong Li
Yuxiong He
MoE
160
413
0
18 Jan 2021
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
243
1,817
0
17 Sep 2019
Linear Convergence of Gradient and Proximal-Gradient Methods Under the Polyak-Łojasiewicz Condition
Hamed Karimi
J. Nutini
Mark W. Schmidt
119
1,198
0
16 Aug 2016
1