ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2008.13006
  4. Cited By
Accelerating Sparse DNN Models without Hardware-Support via Tile-Wise
  Sparsity

Accelerating Sparse DNN Models without Hardware-Support via Tile-Wise Sparsity

29 August 2020
Cong Guo
B. Hsueh
Jingwen Leng
Yuxian Qiu
Yue Guan
Zehuan Wang
Xiaoying Jia
Xipeng Li
Minyi Guo
Yuhao Zhu
ArXiv (abs)PDFHTML

Papers citing "Accelerating Sparse DNN Models without Hardware-Support via Tile-Wise Sparsity"

26 / 26 papers shown
Title
Phi: Leveraging Pattern-based Hierarchical Sparsity for High-Efficiency Spiking Neural Networks
Phi: Leveraging Pattern-based Hierarchical Sparsity for High-Efficiency Spiking Neural Networks
Chiyue Wei
Bowen Duan
Cong Guo
Jing Zhang
Qingyue Song
Hai "Helen" Li
Yiran Chen
152
0
0
16 May 2025
Accelerating Sparse DNNs Based on Tiled GEMM
Accelerating Sparse DNNs Based on Tiled GEMM
Cong Guo
Fengchen Xue
Jingwen Leng
Yuxian Qiu
Yue Guan
Weihao Cui
Quan Chen
Minyi Guo
80
13
0
16 Feb 2024
AdaptGear: Accelerating GNN Training via Adaptive Subgraph-Level Kernels
  on GPUs
AdaptGear: Accelerating GNN Training via Adaptive Subgraph-Level Kernels on GPUs
Yangjie Zhou
Yaoxu Song
Jingwen Leng
Zihan Liu
Weihao Cui
Zhendong Zhang
Cong Guo
Quan Chen
Li-Wei Li
Minyi Guo
GNN
109
3
0
27 May 2023
Energy-Latency Attacks to On-Device Neural Networks via Sponge Poisoning
Energy-Latency Attacks to On-Device Neural Networks via Sponge Poisoning
Zijian Wang
Shuo Huang
Yu-Jen Huang
Helei Cui
SILM
95
12
0
06 May 2023
Gradient-Free Structured Pruning with Unlabeled Data
Gradient-Free Structured Pruning with Unlabeled Data
Azade Nova
H. Dai
Dale Schuurmans
SyDa
158
28
0
07 Mar 2023
TensorFHE: Achieving Practical Computation on Encrypted Data Using GPGPU
TensorFHE: Achieving Practical Computation on Encrypted Data Using GPGPU
Shengyu Fan
Zhiwei Wang
Weizhi Xu
Rui Hou
Dan Meng
Hao Fei
FedML
93
56
0
29 Dec 2022
LearningGroup: A Real-Time Sparse Training on FPGA via Learnable Weight
  Grouping for Multi-Agent Reinforcement Learning
LearningGroup: A Real-Time Sparse Training on FPGA via Learnable Weight Grouping for Multi-Agent Reinforcement Learning
Jenny Yang
Jaeuk Kim
Joo-Young Kim
95
2
0
29 Oct 2022
Nesting Forward Automatic Differentiation for Memory-Efficient Deep
  Neural Network Training
Nesting Forward Automatic Differentiation for Memory-Efficient Deep Neural Network Training
Cong Guo
Yuxian Qiu
Jingwen Leng
Chen Zhang
Yingdian Cao
Quan Zhang
Yunxin Liu
Fan Yang
Minyi Guo
AI4CE
117
4
0
22 Sep 2022
ANT: Exploiting Adaptive Numerical Data Type for Low-bit Deep Neural
  Network Quantization
ANT: Exploiting Adaptive Numerical Data Type for Low-bit Deep Neural Network Quantization
Cong Guo
Chen Zhang
Jingwen Leng
Zihan Liu
Fan Yang
Yun-Bo Liu
Minyi Guo
Yuhao Zhu
MQ
122
72
0
30 Aug 2022
A One-Shot Reparameterization Method for Reducing the Loss of Tile
  Pruning on DNNs
A One-Shot Reparameterization Method for Reducing the Loss of Tile Pruning on DNNs
Yancheng Li
Qingzhong Ai
Fumihiko Ino
99
0
0
29 Jul 2022
Monarch: Expressive Structured Matrices for Efficient and Accurate
  Training
Monarch: Expressive Structured Matrices for Efficient and Accurate Training
Tri Dao
Beidi Chen
N. Sohoni
Arjun D Desai
Michael Poli
Jessica Grogan
Alexander Liu
Aniruddh Rao
Atri Rudra
Christopher Ré
177
101
0
01 Apr 2022
A Fast Post-Training Pruning Framework for Transformers
A Fast Post-Training Pruning Framework for Transformers
Woosuk Kwon
Sehoon Kim
Michael W. Mahoney
Joseph Hassoun
Kurt Keutzer
A. Gholami
154
175
0
29 Mar 2022
Shfl-BW: Accelerating Deep Neural Network Inference with Tensor-Core
  Aware Weight Pruning
Shfl-BW: Accelerating Deep Neural Network Inference with Tensor-Core Aware Weight Pruning
Guyue Huang
Haoran Li
Minghai Qin
Fei Sun
Yufei Din
Yuan Xie
116
19
0
09 Mar 2022
SQuant: On-the-Fly Data-Free Quantization via Diagonal Hessian
  Approximation
SQuant: On-the-Fly Data-Free Quantization via Diagonal Hessian Approximation
Cong Guo
Yuxian Qiu
Jingwen Leng
Xiaotian Gao
Chen Zhang
Yunxin Liu
Fan Yang
Yuhao Zhu
Minyi Guo
MQ
164
80
0
14 Feb 2022
VELTAIR: Towards High-Performance Multi-tenant Deep Learning Services
  via Adaptive Compilation and Scheduling
VELTAIR: Towards High-Performance Multi-tenant Deep Learning Services via Adaptive Compilation and Scheduling
Zihan Liu
Jingwen Leng
Zhihui Zhang
Quan Chen
Chao Li
Minyi Guo
80
50
0
17 Jan 2022
Extending the limit of molecular dynamics with ab initio accuracy to 10
  billion atoms
Extending the limit of molecular dynamics with ab initio accuracy to 10 billion atoms
Zhuoqiang Guo
Denghui Lu
Yujin Yan
Siyu Hu
Rongrong Liu
...
Yixiao Chen
Linfeng Zhang
Mohan Chen
Han Wang
Weile Jia
AI4CE
91
47
0
05 Jan 2022
SPViT: Enabling Faster Vision Transformers via Soft Token Pruning
SPViT: Enabling Faster Vision Transformers via Soft Token Pruning
Zhenglun Kong
Zhaoyang Han
Xiaolong Ma
Xin Meng
Mengshu Sun
...
Geng Yuan
Bin Ren
Minghai Qin
Hao Tang
Yanzhi Wang
ViT
143
171
0
27 Dec 2021
Block-Skim: Efficient Question Answering for Transformer
Block-Skim: Efficient Question Answering for Transformer
Yue Guan
Zhengyi Li
Jingwen Leng
Zhouhan Lin
Minyi Guo
Yuhao Zhu
97
32
0
16 Dec 2021
Accelerating Framework of Transformer by Hardware Design and Model
  Compression Co-Optimization
Accelerating Framework of Transformer by Hardware Design and Model Compression Co-Optimization
Panjie Qi
E. Sha
Qingfeng Zhuge
Hongwu Peng
Shaoyi Huang
Zhenglun Kong
Yuhong Song
Bingbing Li
111
52
0
19 Oct 2021
Characterizing and Demystifying the Implicit Convolution Algorithm on
  Commercial Matrix-Multiplication Accelerators
Characterizing and Demystifying the Implicit Convolution Algorithm on Commercial Matrix-Multiplication Accelerators
Yangjie Zhou
Mengtian Yang
Cong Guo
Jingwen Leng
Yun Liang
Quan Chen
Minyi Guo
Yuhao Zhu
79
36
0
08 Oct 2021
RED++ : Data-Free Pruning of Deep Neural Networks via Input Splitting
  and Output Merging
RED++ : Data-Free Pruning of Deep Neural Networks via Input Splitting and Output Merging
Edouard Yvinec
Arnaud Dapogny
Matthieu Cord
Kévin Bailly
143
22
0
30 Sep 2021
Dual-side Sparse Tensor Core
Dual-side Sparse Tensor Core
Yang-Feng Wang
Chen Zhang
Zhiqiang Xie
Cong Guo
Yunxin Liu
Jingwen Leng
121
83
0
20 May 2021
Post-training deep neural network pruning via layer-wise calibration
Post-training deep neural network pruning via layer-wise calibration
Ivan Lazarevich
Alexander Kozlov
Nikita Malinin
3DPC
95
33
0
30 Apr 2021
Partitioning sparse deep neural networks for scalable training and
  inference
Partitioning sparse deep neural networks for scalable training and inference
G. Demirci
Hakan Ferhatosmanoglu
80
12
0
23 Apr 2021
GPU Semiring Primitives for Sparse Neighborhood Methods
GPU Semiring Primitives for Sparse Neighborhood Methods
Corey J. Nolet
Divye Gala
Edward Raff
Joe Eaton
Brad Rees
John Zedlewski
Tim Oates
127
6
0
13 Apr 2021
DLFusion: An Auto-Tuning Compiler for Layer Fusion on Deep Neural
  Network Accelerator
DLFusion: An Auto-Tuning Compiler for Layer Fusion on Deep Neural Network Accelerator
Zihan Liu
Jingwen Leng
Quan Chen
Chao Li
Wenli Zheng
Li-Wei Li
Minyi Guo
55
8
0
11 Nov 2020
1