Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2008.11421
Cited By
Scaling Distributed Deep Learning Workloads beyond the Memory Capacity with KARMA
26 August 2020
M. Wahib
Haoyu Zhang
Truong Thao Nguyen
Aleksandr Drozd
Jens Domke
Lingqi Zhang
Ryousei Takano
Satoshi Matsuoka
OODD
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Scaling Distributed Deep Learning Workloads beyond the Memory Capacity with KARMA"
2 / 2 papers shown
Title
Democratizing AI: Open-source Scalable LLM Training on GPU-based Supercomputers
Siddharth Singh
Prajwal Singhania
Aditya K. Ranjan
John Kirchenbauer
Jonas Geiping
...
Abhimanyu Hans
Manli Shu
Aditya Tomar
Tom Goldstein
A. Bhatele
92
2
0
12 Feb 2025
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
243
1,791
0
17 Sep 2019
1