ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2209.12769
  4. Cited By
Optimizing DNN Compilation for Distributed Training with Joint OP and
  Tensor Fusion

Optimizing DNN Compilation for Distributed Training with Joint OP and Tensor Fusion

26 September 2022
Xiaodong Yi
Shiwei Zhang
Lansong Diao
Chuan Wu
Zhen Zheng
Shiqing Fan
Siyu Wang
Jun Yang
W. Lin
ArXivPDFHTML

Papers citing "Optimizing DNN Compilation for Distributed Training with Joint OP and Tensor Fusion"

3 / 3 papers shown
Title
TiMePReSt: Time and Memory Efficient Pipeline Parallel DNN Training with
  Removed Staleness
TiMePReSt: Time and Memory Efficient Pipeline Parallel DNN Training with Removed Staleness
Ankita Dutta
Nabendu Chaki
Rajat K. De
27
0
0
18 Oct 2024
Megatron-LM: Training Multi-Billion Parameter Language Models Using
  Model Parallelism
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
245
1,817
0
17 Sep 2019
Deep Graph Library: A Graph-Centric, Highly-Performant Package for Graph
  Neural Networks
Deep Graph Library: A Graph-Centric, Highly-Performant Package for Graph Neural Networks
Minjie Wang
Da Zheng
Zihao Ye
Quan Gan
Mufei Li
...
J. Zhao
Haotong Zhang
Alex Smola
Jinyang Li
Zheng-Wei Zhang
AI4CE
GNN
194
745
0
03 Sep 2019
1