ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1901.09504
  4. Cited By
Improving Neural Network Quantization without Retraining using Outlier
  Channel Splitting

Improving Neural Network Quantization without Retraining using Outlier Channel Splitting

28 January 2019
Ritchie Zhao
Yuwei Hu
Jordan Dotzel
Christopher De Sa
Zhiru Zhang
    OODD
    MQ
ArXivPDFHTML

Papers citing "Improving Neural Network Quantization without Retraining using Outlier Channel Splitting"

24 / 174 papers shown
Title
LSQ+: Improving low-bit quantization through learnable offsets and
  better initialization
LSQ+: Improving low-bit quantization through learnable offsets and better initialization
Yash Bhalgat
Jinwon Lee
Markus Nagel
Tijmen Blankevoort
Nojun Kwak
MQ
17
212
0
20 Apr 2020
From Quantized DNNs to Quantizable DNNs
From Quantized DNNs to Quantizable DNNs
Kunyuan Du
Ya-Qin Zhang
Haibing Guan
MQ
8
3
0
11 Apr 2020
Generative Low-bitwidth Data Free Quantization
Generative Low-bitwidth Data Free Quantization
Shoukai Xu
Haokun Li
Bohan Zhuang
Jing Liu
Jiezhang Cao
Chuangrun Liang
Mingkui Tan
MQ
13
126
0
07 Mar 2020
Quantized Neural Network Inference with Precision Batching
Quantized Neural Network Inference with Precision Batching
Maximilian Lam
Zachary Yedidia
Colby R. Banbury
Vijay Janapa Reddi
MQ
20
1
0
26 Feb 2020
Post-training Quantization with Multiple Points: Mixed Precision without
  Mixed Precision
Post-training Quantization with Multiple Points: Mixed Precision without Mixed Precision
Xingchao Liu
Mao Ye
Dengyong Zhou
Qiang Liu
MQ
8
42
0
20 Feb 2020
Robust Quantization: One Model to Rule Them All
Robust Quantization: One Model to Rule Them All
Moran Shkolnik
Brian Chmiel
Ron Banner
Gil Shomron
Yury Nahshan
A. Bronstein
U. Weiser
OOD
MQ
6
75
0
18 Feb 2020
Gradient $\ell_1$ Regularization for Quantization Robustness
Gradient ℓ1\ell_1ℓ1​ Regularization for Quantization Robustness
Milad Alizadeh
Arash Behboodi
M. V. Baalen
Christos Louizos
Tijmen Blankevoort
Max Welling
MQ
12
8
0
18 Feb 2020
Precision Gating: Improving Neural Network Efficiency with Dynamic
  Dual-Precision Activations
Precision Gating: Improving Neural Network Efficiency with Dynamic Dual-Precision Activations
Yichi Zhang
Ritchie Zhao
Weizhe Hua
N. Xu
G. E. Suh
Zhiru Zhang
MQ
82
27
0
17 Feb 2020
Post-Training Piecewise Linear Quantization for Deep Neural Networks
Post-Training Piecewise Linear Quantization for Deep Neural Networks
Jun Fang
Ali Shafiee
Hamzah Abdel-Aziz
D. Thorsley
Georgios Georgiadis
Joseph Hassoun
MQ
10
143
0
31 Jan 2020
RPR: Random Partition Relaxation for Training; Binary and Ternary Weight
  Neural Networks
RPR: Random Partition Relaxation for Training; Binary and Ternary Weight Neural Networks
Lukas Cavigelli
Luca Benini
MQ
13
9
0
04 Jan 2020
ZeroQ: A Novel Zero Shot Quantization Framework
ZeroQ: A Novel Zero Shot Quantization Framework
Yaohui Cai
Z. Yao
Zhen Dong
A. Gholami
Michael W. Mahoney
Kurt Keutzer
MQ
30
389
0
01 Jan 2020
WaLDORf: Wasteless Language-model Distillation On Reading-comprehension
WaLDORf: Wasteless Language-model Distillation On Reading-comprehension
J. Tian
A. Kreuzer
Pai-Hung Chen
Hans-Martin Will
VLM
34
3
0
13 Dec 2019
Loss Aware Post-training Quantization
Loss Aware Post-training Quantization
Yury Nahshan
Brian Chmiel
Chaim Baskin
Evgenii Zheltonozhskii
Ron Banner
A. Bronstein
A. Mendelson
MQ
17
163
0
17 Nov 2019
HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks
HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks
Zhen Dong
Z. Yao
Yaohui Cai
Daiyaan Arfeen
A. Gholami
Michael W. Mahoney
Kurt Keutzer
MQ
26
274
0
10 Nov 2019
Post-Training 4-bit Quantization on Embedding Tables
Post-Training 4-bit Quantization on Embedding Tables
Hui Guan
Andrey Malevich
Jiyan Yang
Jongsoo Park
Hector Yuen
MQ
6
31
0
05 Nov 2019
Neural Network Distiller: A Python Package For DNN Compression Research
Neural Network Distiller: A Python Package For DNN Compression Research
Neta Zmora
Guy Jacob
Lev Zlotnik
Bar Elharar
Gal Novik
17
73
0
27 Oct 2019
OverQ: Opportunistic Outlier Quantization for Neural Network
  Accelerators
OverQ: Opportunistic Outlier Quantization for Neural Network Accelerators
Ritchie Zhao
Jordan Dotzel
Zhanqiu Hu
Preslav Ivanov
Christopher De Sa
Zhiru Zhang
MQ
14
1
0
13 Oct 2019
QuaRL: Quantization for Fast and Environmentally Sustainable
  Reinforcement Learning
QuaRL: Quantization for Fast and Environmentally Sustainable Reinforcement Learning
Srivatsan Krishnan
Maximilian Lam
Sharad Chitlangia
Zishen Wan
Gabriel Barth-Maron
Aleksandra Faust
Vijay Janapa Reddi
MQ
15
22
0
02 Oct 2019
Additive Powers-of-Two Quantization: An Efficient Non-uniform
  Discretization for Neural Networks
Additive Powers-of-Two Quantization: An Efficient Non-uniform Discretization for Neural Networks
Yuhang Li
Xin Dong
Wei Wang
MQ
15
254
0
28 Sep 2019
An Inter-Layer Weight Prediction and Quantization for Deep Neural
  Networks based on a Smoothly Varying Weight Hypothesis
An Inter-Layer Weight Prediction and Quantization for Deep Neural Networks based on a Smoothly Varying Weight Hypothesis
Kang-Ho Lee
Joonhyun Jeong
Sung-Ho Bae
22
4
0
16 Jul 2019
Fighting Quantization Bias With Bias
Fighting Quantization Bias With Bias
Alexander Finkelstein
Uri Almog
Mark Grobman
MQ
12
56
0
07 Jun 2019
Instant Quantization of Neural Networks using Monte Carlo Methods
Instant Quantization of Neural Networks using Monte Carlo Methods
Gonçalo Mordido
Matthijs Van Keirsbilck
A. Keller
MQ
16
9
0
29 May 2019
DSConv: Efficient Convolution Operator
DSConv: Efficient Convolution Operator
Marcelo Gennari
Roger Fawcett
V. Prisacariu
MQ
24
62
0
07 Jan 2019
Incremental Network Quantization: Towards Lossless CNNs with
  Low-Precision Weights
Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights
Aojun Zhou
Anbang Yao
Yiwen Guo
Lin Xu
Yurong Chen
MQ
311
1,047
0
10 Feb 2017
Previous
1234