Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2001.03288
Cited By
Efficient Memory Management for Deep Neural Net Inference
10 January 2020
Yury Pisarchyk
Juhyun Lee
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Efficient Memory Management for Deep Neural Net Inference"
5 / 5 papers shown
Title
Scaling On-Device GPU Inference for Large Generative Models
Jiuqiang Tang
Raman Sarokin
Ekaterina Ignasheva
Grant Jensen
Lin Chen
Juhyun Lee
Andrei Kulik
Matthias Grundmann
118
1
0
01 May 2025
Data Overfitting for On-Device Super-Resolution with Dynamic Algorithm and Compiler Co-Design
Gen Li
Zhihao Shu
Jie Ji
Minghai Qin
Fatemeh Afghah
Wei Niu
Xiaolong Ma
SupR
38
0
0
03 Jul 2024
Demand Layering for Real-Time DNN Inference with Minimized Memory Usage
Min-Zhi Ji
Saehanseul Yi
Chang-Mo Koo
Sol Ahn
Dongjoo Seo
N. Dutt
Jong-Chan Kim
42
16
0
08 Oct 2022
Memory Planning for Deep Neural Networks
Maksim Levental
23
4
0
23 Feb 2022
PatrickStar: Parallel Training of Pre-trained Models via Chunk-based Memory Management
Jiarui Fang
Zilin Zhu
Shenggui Li
Hui Su
Yang Yu
Jie Zhou
Yang You
VLM
29
24
0
12 Aug 2021
1