Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2404.10933
Cited By
LLMem: Estimating GPU Memory Usage for Fine-Tuning Pre-Trained LLMs
16 April 2024
Taeho Kim
Yanming Wang
Vatshank Chaturvedi
Lokesh Gupta
Seyeon Kim
Yongin Kwon
Sangtae Ha
Re-assign community
ArXiv
PDF
HTML
Papers citing
"LLMem: Estimating GPU Memory Usage for Fine-Tuning Pre-Trained LLMs"
5 / 5 papers shown
Title
Accurate GPU Memory Prediction for Deep Learning Jobs through Dynamic Analysis
Jiabo Shi
Yehia Elkhatib
3DH
VLM
25
0
0
04 Apr 2025
KD-LoRA: A Hybrid Approach to Efficient Fine-Tuning with LoRA and Knowledge Distillation
Rambod Azimi
Rishav Rishav
M. Teichmann
Samira Ebrahimi Kahou
ALM
18
0
0
28 Oct 2024
ZeRO-Offload: Democratizing Billion-Scale Model Training
Jie Ren
Samyam Rajbhandari
Reza Yazdani Aminabadi
Olatunji Ruwase
Shuangyang Yang
Minjia Zhang
Dong Li
Yuxiong He
MoE
160
399
0
18 Jan 2021
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
243
1,791
0
17 Sep 2019
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
294
6,927
0
20 Apr 2018
1