ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2008.11421
  4. Cited By
Scaling Distributed Deep Learning Workloads beyond the Memory Capacity
  with KARMA

Scaling Distributed Deep Learning Workloads beyond the Memory Capacity with KARMA

International Conference for High Performance Computing, Networking, Storage and Analysis (SC), 2020
26 August 2020
Mohamed Wahib
Haoyu Zhang
Truong Thao Nguyen
Aleksandr Drozd
Jens Domke
Lingqi Zhang
Ryousei Takano
Satoshi Matsuoka
    OODD
ArXiv (abs)PDFHTML

Papers citing "Scaling Distributed Deep Learning Workloads beyond the Memory Capacity with KARMA"

5 / 5 papers shown
Democratizing AI: Open-source Scalable LLM Training on GPU-based Supercomputers
Democratizing AI: Open-source Scalable LLM Training on GPU-based SupercomputersInternational Conference for High Performance Computing, Networking, Storage and Analysis (SC), 2024
Siddharth Singh
Prajwal Singhania
Aditya K. Ranjan
John Kirchenbauer
Jonas Geiping
...
Abhimanyu Hans
Manli Shu
Aditya Tomar
Tom Goldstein
A. Bhatele
875
11
0
12 Feb 2025
FedDCT: Federated Learning of Large Convolutional Neural Networks on
  Resource Constrained Devices using Divide and Collaborative Training
FedDCT: Federated Learning of Large Convolutional Neural Networks on Resource Constrained Devices using Divide and Collaborative TrainingIEEE Transactions on Network and Service Management (IEEE TNSM), 2022
Quan Nguyen
Hieu H. Pham
Kok-Seng Wong
Phi Le Nguyen
Truong Thao Nguyen
Minh N. Do
FedML
298
10
0
20 Nov 2022
PERKS: a Locality-Optimized Execution Model for Iterative Memory-bound
  GPU Applications
PERKS: a Locality-Optimized Execution Model for Iterative Memory-bound GPU ApplicationsInternational Conference on Supercomputing (ICS), 2022
Lingqi Zhang
Mohamed Wahib
Peng Chen
Jintao Meng
Xiao Wang
Toshio Endo
Satoshi Matsuoka
270
14
0
05 Apr 2022
A Survey and Empirical Evaluation of Parallel Deep Learning Frameworks
A Survey and Empirical Evaluation of Parallel Deep Learning Frameworks
Daniel Nichols
Siddharth Singh
Shuqing Lin
A. Bhatele
OOD
240
11
0
09 Nov 2021
An Oracle for Guiding Large-Scale Model/Hybrid Parallel Training of
  Convolutional Neural Networks
An Oracle for Guiding Large-Scale Model/Hybrid Parallel Training of Convolutional Neural NetworksIEEE International Symposium on High-Performance Parallel Distributed Computing (HPDC), 2020
A. Kahira
Truong Thao Nguyen
L. Bautista-Gomez
Ryousei Takano
Rosa M. Badia
Mohamed Wahib
199
14
0
19 Apr 2021
1
Page 1 of 1