ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2001.03288
  4. Cited By
Efficient Memory Management for Deep Neural Net Inference

Efficient Memory Management for Deep Neural Net Inference

10 January 2020
Yury Pisarchyk
Juhyun Lee
ArXivPDFHTML

Papers citing "Efficient Memory Management for Deep Neural Net Inference"

5 / 5 papers shown
Title
Scaling On-Device GPU Inference for Large Generative Models
Scaling On-Device GPU Inference for Large Generative Models
Jiuqiang Tang
Raman Sarokin
Ekaterina Ignasheva
Grant Jensen
Lin Chen
Juhyun Lee
Andrei Kulik
Matthias Grundmann
118
1
0
01 May 2025
Data Overfitting for On-Device Super-Resolution with Dynamic Algorithm
  and Compiler Co-Design
Data Overfitting for On-Device Super-Resolution with Dynamic Algorithm and Compiler Co-Design
Gen Li
Zhihao Shu
Jie Ji
Minghai Qin
Fatemeh Afghah
Wei Niu
Xiaolong Ma
SupR
38
0
0
03 Jul 2024
Demand Layering for Real-Time DNN Inference with Minimized Memory Usage
Demand Layering for Real-Time DNN Inference with Minimized Memory Usage
Min-Zhi Ji
Saehanseul Yi
Chang-Mo Koo
Sol Ahn
Dongjoo Seo
N. Dutt
Jong-Chan Kim
42
16
0
08 Oct 2022
Memory Planning for Deep Neural Networks
Memory Planning for Deep Neural Networks
Maksim Levental
23
4
0
23 Feb 2022
PatrickStar: Parallel Training of Pre-trained Models via Chunk-based
  Memory Management
PatrickStar: Parallel Training of Pre-trained Models via Chunk-based Memory Management
Jiarui Fang
Zilin Zhu
Shenggui Li
Hui Su
Yang Yu
Jie Zhou
Yang You
VLM
29
24
0
12 Aug 2021
1