Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1707.09870
Cited By
Extremely Low Bit Neural Network: Squeeze the Last Bit Out with ADMM
24 July 2017
Cong Leng
Hao Li
Shenghuo Zhu
R. L. Jin
MQ
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Extremely Low Bit Neural Network: Squeeze the Last Bit Out with ADMM"
41 / 41 papers shown
Title
Efficient Continual Learning in Keyword Spotting using Binary Neural Networks
Quynh Nguyen Phuong Vu
Luciano S. Martinez-Rau
Yuxuan Zhang
Nho-Duc Tran
Bengt Oelmann
Michele Magno
Sebastian Bader
CLL
38
0
0
05 May 2025
Quality Scalable Quantization Methodology for Deep Learning on Edge
S. Khaliq
Rehan Hafiz
MQ
33
1
0
15 Jul 2024
Quantization Aware Factorization for Deep Neural Network Compression
Daria Cherniuk
Stanislav Abukhovich
Anh-Huy Phan
Ivan V. Oseledets
A. Cichocki
Julia Gusak
MQ
15
2
0
08 Aug 2023
AutoQNN: An End-to-End Framework for Automatically Quantizing Neural Networks
Cheng Gong
Ye Lu
Surong Dai
Deng Qian
Chenkun Du
Tao Li
MQ
27
0
0
07 Apr 2023
Learning Discretized Neural Networks under Ricci Flow
Jun Chen
Han Chen
Mengmeng Wang
Guang Dai
Ivor W. Tsang
Y. Liu
13
2
0
07 Feb 2023
Hyperspherical Quantization: Toward Smaller and More Accurate Models
Dan Liu
X. Chen
Chen-li Ma
Xue Liu
MQ
22
3
0
24 Dec 2022
AskewSGD : An Annealed interval-constrained Optimisation method to train Quantized Neural Networks
Louis Leconte
S. Schechtman
Eric Moulines
27
4
0
07 Nov 2022
Seeking Interpretability and Explainability in Binary Activated Neural Networks
Benjamin Leblanc
Pascal Germain
FAtt
27
1
0
07 Sep 2022
Mixed-Precision Neural Networks: A Survey
M. Rakka
M. Fouda
Pramod P. Khargonekar
Fadi J. Kurdahi
MQ
18
11
0
11 Aug 2022
Deep Neural Networks pruning via the Structured Perspective Regularization
M. Cacciola
A. Frangioni
Xinlin Li
Andrea Lodi
3DPC
28
5
0
28 Jun 2022
DNN Training Acceleration via Exploring GPGPU Friendly Sparsity
Zhuoran Song
Yihong Xu
Han Li
Naifeng Jing
Xiaoyao Liang
Li Jiang
27
3
0
11 Mar 2022
Vau da muntanialas: Energy-efficient multi-die scalable acceleration of RNN inference
G. Paulin
Francesco Conti
Lukas Cavigelli
Luca Benini
20
8
0
14 Feb 2022
Sub-mW Keyword Spotting on an MCU: Analog Binary Feature Extraction and Binary Neural Networks
G. Cerutti
Lukas Cavigelli
Renzo Andri
Michele Magno
Elisabetta Farella
Luca Benini
26
14
0
10 Jan 2022
Mixed Precision of Quantization of Transformer Language Models for Speech Recognition
Junhao Xu
Shoukang Hu
Jianwei Yu
Xunying Liu
Helen M. Meng
MQ
30
15
0
29 Nov 2021
CBP: Backpropagation with constraint on weight precision using a pseudo-Lagrange multiplier method
Guhyun Kim
D. Jeong
MQ
34
2
0
06 Oct 2021
Elastic Significant Bit Quantization and Acceleration for Deep Neural Networks
Cheng Gong
Ye Lu
Kunpeng Xie
Zongming Jin
Tao Li
Yanzhi Wang
MQ
17
7
0
08 Sep 2021
Architecture Aware Latency Constrained Sparse Neural Networks
Tianli Zhao
Qinghao Hu
Xiangyu He
Weixiang Xu
Jiaxing Wang
Cong Leng
Jian Cheng
25
0
0
01 Sep 2021
QuPeD: Quantized Personalization via Distillation with Applications to Federated Learning
Kaan Ozkara
Navjot Singh
Deepesh Data
Suhas Diggavi
FedML
MQ
24
56
0
29 Jul 2021
3U-EdgeAI: Ultra-Low Memory Training, Ultra-Low BitwidthQuantization, and Ultra-Low Latency Acceleration
Yao Chen
Cole Hawkins
Kaiqi Zhang
Zheng-Wei Zhang
Cong Hao
11
8
0
11 May 2021
Mix and Match: A Novel FPGA-Centric Deep Neural Network Quantization Framework
Sung-En Chang
Yanyu Li
Mengshu Sun
Runbin Shi
Hayden Kwok-Hay So
Xuehai Qian
Yanzhi Wang
Xue Lin
MQ
18
82
0
08 Dec 2020
MSP: An FPGA-Specific Mixed-Scheme, Multi-Precision Deep Neural Network Quantization Framework
Sung-En Chang
Yanyu Li
Mengshu Sun
Weiwen Jiang
Runbin Shi
Xue Lin
Yanzhi Wang
MQ
11
7
0
16 Sep 2020
Automatic low-bit hybrid quantization of neural networks through meta learning
Tao Wang
Junsong Wang
Chang Xu
Chao Xue
MQ
6
2
0
24 Apr 2020
Binary Neural Networks: A Survey
Haotong Qin
Ruihao Gong
Xianglong Liu
Xiao Bai
Jingkuan Song
N. Sebe
MQ
34
457
0
31 Mar 2020
LANCE: Efficient Low-Precision Quantized Winograd Convolution for Neural Networks Based on Graphics Processing Units
Guangli Li
Lei Liu
Xueying Wang
Xiu Ma
Xiaobing Feng
MQ
11
18
0
19 Mar 2020
Compact recurrent neural networks for acoustic event detection on low-energy low-complexity platforms
G. Cerutti
Rahul Prasad
A. Brutti
Elisabetta Farella
13
47
0
29 Jan 2020
Towards Efficient Training for Neural Network Quantization
Qing Jin
Linjie Yang
Zhenyu A. Liao
MQ
11
42
0
21 Dec 2019
Quantization Networks
Jiwei Yang
Xu Shen
Jun Xing
Xinmei Tian
Houqiang Li
Bing Deng
Jianqiang Huang
Xiansheng Hua
MQ
20
338
0
21 Nov 2019
Bayesian Optimized 1-Bit CNNs
Jiaxin Gu
Junhe Zhao
Xiaolong Jiang
Baochang Zhang
Jianzhuang Liu
G. Guo
Rongrong Ji
BDL
MQ
14
56
0
17 Aug 2019
Weight Normalization based Quantization for Deep Neural Network Compression
Wenhong Cai
Wu-Jun Li
16
14
0
01 Jul 2019
Toward Extremely Low Bit and Lossless Accuracy in DNNs with Progressive ADMM
Sheng Lin
Xiaolong Ma
Shaokai Ye
Geng Yuan
Kaisheng Ma
Yanzhi Wang
MQ
17
10
0
02 May 2019
Progressive DNN Compression: A Key to Achieve Ultra-High Weight Pruning and Quantization Rates using ADMM
Shaokai Ye
Xiaoyu Feng
Tianyun Zhang
Xiaolong Ma
Sheng Lin
...
Jian Tang
M. Fardad
X. Lin
Yongpan Liu
Yanzhi Wang
MQ
22
38
0
23 Mar 2019
Focused Quantization for Sparse CNNs
Yiren Zhao
Xitong Gao
Daniel Bates
Robert D. Mullins
Chengzhong Xu
MQ
15
26
0
07 Mar 2019
Learning low-precision neural networks without Straight-Through Estimator(STE)
Z. G. Liu
Matthew Mattina
MQ
11
34
0
04 Mar 2019
Progressive Weight Pruning of Deep Neural Networks using ADMM
Shaokai Ye
Tianyun Zhang
Kaiqi Zhang
Jiayu Li
Kaidi Xu
...
M. Fardad
Sijia Liu
Xiang Chen
X. Lin
Yanzhi Wang
AI4CE
21
38
0
17 Oct 2018
Quantization for Rapid Deployment of Deep Neural Networks
J. Lee
Sangwon Ha
Saerom Choi
Won-Jo Lee
Seungwon Lee
MQ
11
48
0
12 Oct 2018
Simultaneously Optimizing Weight and Quantizer of Ternary Neural Network using Truncated Gaussian Approximation
Zhezhi He
Deliang Fan
MQ
9
66
0
02 Oct 2018
A Survey on Methods and Theories of Quantized Neural Networks
Yunhui Guo
MQ
27
230
0
13 Aug 2018
Loss-aware Weight Quantization of Deep Networks
Lu Hou
James T. Kwok
MQ
13
127
0
23 Feb 2018
BinaryRelax: A Relaxation Approach For Training Deep Neural Networks With Quantized Weights
Penghang Yin
Shuai Zhang
J. Lyu
Stanley Osher
Y. Qi
Jack Xin
MQ
22
78
0
19 Jan 2018
Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference
Benoit Jacob
S. Kligys
Bo Chen
Menglong Zhu
Matthew Tang
Andrew G. Howard
Hartwig Adam
Dmitry Kalenichenko
MQ
16
3,043
0
15 Dec 2017
Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights
Aojun Zhou
Anbang Yao
Yiwen Guo
Lin Xu
Yurong Chen
MQ
311
1,047
0
10 Feb 2017
1