ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1705.08922
  4. Cited By
Exploring the Regularity of Sparse Structure in Convolutional Neural
  Networks

Exploring the Regularity of Sparse Structure in Convolutional Neural Networks

24 May 2017
Huizi Mao
Song Han
Jeff Pool
Wenshuo Li
Xingyu Liu
Yu Wang
W. Dally
ArXivPDFHTML

Papers citing "Exploring the Regularity of Sparse Structure in Convolutional Neural Networks"

32 / 32 papers shown
Title
Training Acceleration of Low-Rank Decomposed Networks using Sequential
  Freezing and Rank Quantization
Training Acceleration of Low-Rank Decomposed Networks using Sequential Freezing and Rank Quantization
Habib Hajimolahoseini
Walid Ahmed
Yang Liu
OffRL
MQ
19
6
0
07 Sep 2023
Efficient Quantized Sparse Matrix Operations on Tensor Cores
Efficient Quantized Sparse Matrix Operations on Tensor Cores
Shigang Li
Kazuki Osawa
Torsten Hoefler
72
31
0
14 Sep 2022
Towards Sparsification of Graph Neural Networks
Towards Sparsification of Graph Neural Networks
Hongwu Peng
Deniz Gurevin
Shaoyi Huang
Tong Geng
Weiwen Jiang
O. Khan
Caiwen Ding
GNN
30
24
0
11 Sep 2022
Entropy Induced Pruning Framework for Convolutional Neural Networks
Entropy Induced Pruning Framework for Convolutional Neural Networks
Yihe Lu
Ziyu Guan
Yaming Yang
Maoguo Gong
Wei Zhao
Kaiyuan Feng
22
2
0
13 Aug 2022
Mixed-Precision Neural Networks: A Survey
Mixed-Precision Neural Networks: A Survey
M. Rakka
M. Fouda
Pramod P. Khargonekar
Fadi J. Kurdahi
MQ
18
11
0
11 Aug 2022
Compiler-Aware Neural Architecture Search for On-Mobile Real-time
  Super-Resolution
Compiler-Aware Neural Architecture Search for On-Mobile Real-time Super-Resolution
Yushu Wu
Yifan Gong
Pu Zhao
Yanyu Li
Zheng Zhan
Wei Niu
Hao Tang
Minghai Qin
Bin Ren
Yanzhi Wang
SupR
MQ
27
23
0
25 Jul 2022
ZeroQuant: Efficient and Affordable Post-Training Quantization for
  Large-Scale Transformers
ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers
Z. Yao
Reza Yazdani Aminabadi
Minjia Zhang
Xiaoxia Wu
Conglong Li
Yuxiong He
VLM
MQ
45
440
0
04 Jun 2022
LilNetX: Lightweight Networks with EXtreme Model Compression and
  Structured Sparsification
LilNetX: Lightweight Networks with EXtreme Model Compression and Structured Sparsification
Sharath Girish
Kamal Gupta
Saurabh Singh
Abhinav Shrivastava
28
11
0
06 Apr 2022
Quantization in Layer's Input is Matter
Quantization in Layer's Input is Matter
Daning Cheng
Wenguang Chen
MQ
11
0
0
10 Feb 2022
EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network
  Accelerators
EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators
Lois Orosa
Skanda Koppula
Yaman Umuroglu
Konstantinos Kanellopoulos
Juan Gómez Luna
Michaela Blott
K. Vissers
O. Mutlu
35
4
0
04 Feb 2022
Improving the Accuracy of Early Exits in Multi-Exit Architectures via
  Curriculum Learning
Improving the Accuracy of Early Exits in Multi-Exit Architectures via Curriculum Learning
Arian Bakhtiarnia
Qi Zhang
Alexandros Iosifidis
28
12
0
21 Apr 2021
RingCNN: Exploiting Algebraically-Sparse Ring Tensors for
  Energy-Efficient CNN-Based Computational Imaging
RingCNN: Exploiting Algebraically-Sparse Ring Tensors for Energy-Efficient CNN-Based Computational Imaging
Chao-Tsung Huang
32
10
0
19 Apr 2021
Lottery Jackpots Exist in Pre-trained Models
Lottery Jackpots Exist in Pre-trained Models
Yu-xin Zhang
Mingbao Lin
Yan Wang
Fei Chao
Rongrong Ji
30
15
0
18 Apr 2021
Dancing along Battery: Enabling Transformer with Run-time
  Reconfigurability on Mobile Devices
Dancing along Battery: Enabling Transformer with Run-time Reconfigurability on Mobile Devices
Yuhong Song
Weiwen Jiang
Bingbing Li
Panjie Qi
Qingfeng Zhuge
E. Sha
Sakyasingha Dasgupta
Yiyu Shi
Caiwen Ding
15
18
0
12 Feb 2021
BRDS: An FPGA-based LSTM Accelerator with Row-Balanced Dual-Ratio
  Sparsification
BRDS: An FPGA-based LSTM Accelerator with Row-Balanced Dual-Ratio Sparsification
Seyed Abolfazl Ghasemzadeh
E. Tavakoli
M. Kamal
A. Afzali-Kusha
Massoud Pedram
8
13
0
07 Jan 2021
Parallel Blockwise Knowledge Distillation for Deep Neural Network
  Compression
Parallel Blockwise Knowledge Distillation for Deep Neural Network Compression
Cody Blakeney
Xiaomin Li
Yan Yan
Ziliang Zong
32
39
0
05 Dec 2020
Self-grouping Convolutional Neural Networks
Self-grouping Convolutional Neural Networks
Qingbei Guo
Xiaojun Wu
J. Kittler
Zhiquan Feng
17
22
0
29 Sep 2020
CSB-RNN: A Faster-than-Realtime RNN Acceleration Framework with
  Compressed Structured Blocks
CSB-RNN: A Faster-than-Realtime RNN Acceleration Framework with Compressed Structured Blocks
Runbin Shi
Peiyan Dong
Tong Geng
Yuhao Ding
Xiaolong Ma
Hayden Kwok-Hay So
Martin C. Herbordt
Ang Li
Yanzhi Wang
MQ
10
13
0
11 May 2020
A Survey of Convolutional Neural Networks: Analysis, Applications, and
  Prospects
A Survey of Convolutional Neural Networks: Analysis, Applications, and Prospects
Zewen Li
Wenjie Yang
Shouheng Peng
Fan Liu
HAI
3DV
54
2,595
0
01 Apr 2020
An Image Enhancing Pattern-based Sparsity for Real-time Inference on
  Mobile Devices
An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices
Xiaolong Ma
Wei Niu
Tianyun Zhang
Sijia Liu
Sheng Lin
...
Xiang Chen
Jian Tang
Kaisheng Ma
Bin Ren
Yanzhi Wang
30
27
0
20 Jan 2020
Sparse Weight Activation Training
Sparse Weight Activation Training
Md Aamir Raihan
Tor M. Aamodt
32
72
0
07 Jan 2020
ZeroQ: A Novel Zero Shot Quantization Framework
ZeroQ: A Novel Zero Shot Quantization Framework
Yaohui Cai
Z. Yao
Zhen Dong
A. Gholami
Michael W. Mahoney
Kurt Keutzer
MQ
27
389
0
01 Jan 2020
A Pre-defined Sparse Kernel Based Convolution for Deep CNNs
A Pre-defined Sparse Kernel Based Convolution for Deep CNNs
Souvik Kundu
Saurav Prakash
H. Akrami
P. Beerel
K. Chugg
28
12
0
02 Oct 2019
PCONV: The Missing but Desirable Sparsity in DNN Weight Pruning for
  Real-time Execution on Mobile Devices
PCONV: The Missing but Desirable Sparsity in DNN Weight Pruning for Real-time Execution on Mobile Devices
Xiaolong Ma
Fu-Ming Guo
Wei Niu
Xue Lin
Jian Tang
Kaisheng Ma
Bin Ren
Yanzhi Wang
CVBM
14
173
0
06 Sep 2019
Parameterized Structured Pruning for Deep Neural Networks
Parameterized Structured Pruning for Deep Neural Networks
Günther Schindler
Wolfgang Roth
Franz Pernkopf
Holger Froening
16
6
0
12 Jun 2019
Towards Efficient Model Compression via Learned Global Ranking
Towards Efficient Model Compression via Learned Global Ranking
Ting-Wu Chin
Ruizhou Ding
Cha Zhang
Diana Marculescu
10
170
0
28 Apr 2019
Progressive DNN Compression: A Key to Achieve Ultra-High Weight Pruning
  and Quantization Rates using ADMM
Progressive DNN Compression: A Key to Achieve Ultra-High Weight Pruning and Quantization Rates using ADMM
Shaokai Ye
Xiaoyu Feng
Tianyun Zhang
Xiaolong Ma
Sheng Lin
...
Jian Tang
M. Fardad
X. Lin
Yongpan Liu
Yanzhi Wang
MQ
27
38
0
23 Mar 2019
ADMM-NN: An Algorithm-Hardware Co-Design Framework of DNNs Using
  Alternating Direction Method of Multipliers
ADMM-NN: An Algorithm-Hardware Co-Design Framework of DNNs Using Alternating Direction Method of Multipliers
Ao Ren
Tianyun Zhang
Shaokai Ye
Jiayu Li
Wenyao Xu
Xuehai Qian
X. Lin
Yanzhi Wang
MQ
24
162
0
31 Dec 2018
Exploiting Kernel Sparsity and Entropy for Interpretable CNN Compression
Exploiting Kernel Sparsity and Entropy for Interpretable CNN Compression
Yuchao Li
Shaohui Lin
Baochang Zhang
Jianzhuang Liu
David Doermann
Yongjian Wu
Feiyue Huang
R. Ji
34
130
0
11 Dec 2018
Progressive Weight Pruning of Deep Neural Networks using ADMM
Progressive Weight Pruning of Deep Neural Networks using ADMM
Shaokai Ye
Tianyun Zhang
Kaiqi Zhang
Jiayu Li
Kaidi Xu
...
M. Fardad
Sijia Liu
Xiang Chen
X. Lin
Yanzhi Wang
AI4CE
23
38
0
17 Oct 2018
IGCV$2$: Interleaved Structured Sparse Convolutional Neural Networks
IGCV222: Interleaved Structured Sparse Convolutional Neural Networks
Guotian Xie
Jingdong Wang
Ting Zhang
Jianhuang Lai
Richang Hong
Guo-Jun Qi
8
104
0
17 Apr 2018
Efficient Hardware Realization of Convolutional Neural Networks using
  Intra-Kernel Regular Pruning
Efficient Hardware Realization of Convolutional Neural Networks using Intra-Kernel Regular Pruning
Maurice Yang
Mahmoud Faraj
Assem Hussein
V. Gaudet
CVBM
14
12
0
15 Mar 2018
1