ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1911.07190
  4. Cited By
Loss Aware Post-training Quantization

Loss Aware Post-training Quantization

17 November 2019
Yury Nahshan
Brian Chmiel
Chaim Baskin
Evgenii Zheltonozhskii
Ron Banner
A. Bronstein
A. Mendelson
    MQ
ArXivPDFHTML

Papers citing "Loss Aware Post-training Quantization"

34 / 84 papers shown
Title
Optimal Brain Compression: A Framework for Accurate Post-Training
  Quantization and Pruning
Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning
Elias Frantar
Sidak Pal Singh
Dan Alistarh
MQ
17
214
0
24 Aug 2022
Symmetry Regularization and Saturating Nonlinearity for Robust
  Quantization
Symmetry Regularization and Saturating Nonlinearity for Robust Quantization
Sein Park
Yeongsang Jang
Eunhyeok Park
MQ
14
1
0
31 Jul 2022
Mixed-Precision Inference Quantization: Radically Towards Faster
  inference speed, Lower Storage requirement, and Lower Loss
Mixed-Precision Inference Quantization: Radically Towards Faster inference speed, Lower Storage requirement, and Lower Loss
Daning Cheng
Wenguang Chen
MQ
24
0
0
20 Jul 2022
Attention Round for Post-Training Quantization
Attention Round for Post-Training Quantization
Huabin Diao
Gongyang Li
Shaoyun Xu
Yuexing Hao
MQ
11
10
0
07 Jul 2022
Quantization Robust Federated Learning for Efficient Inference on
  Heterogeneous Devices
Quantization Robust Federated Learning for Efficient Inference on Heterogeneous Devices
Kartik Gupta
Marios Fournarakis
M. Reisser
Christos Louizos
Markus Nagel
FedML
14
14
0
22 Jun 2022
AMED: Automatic Mixed-Precision Quantization for Edge Devices
AMED: Automatic Mixed-Precision Quantization for Edge Devices
Moshe Kimhi
T. Rozen
A. Mendelson
Chaim Baskin
MQ
15
3
0
30 May 2022
Towards Feature Distribution Alignment and Diversity Enhancement for
  Data-Free Quantization
Towards Feature Distribution Alignment and Diversity Enhancement for Data-Free Quantization
Yangcheng Gao
Zhao Zhang
Richang Hong
Haijun Zhang
Jicong Fan
Shuicheng Yan
MQ
12
10
0
30 Apr 2022
It's All In the Teacher: Zero-Shot Quantization Brought Closer to the
  Teacher
It's All In the Teacher: Zero-Shot Quantization Brought Closer to the Teacher
Kanghyun Choi
Hye Yoon Lee
Deokki Hong
Joonsang Yu
Noseong Park
Youngsok Kim
Jinho Lee
MQ
22
31
0
31 Mar 2022
Overcoming Oscillations in Quantization-Aware Training
Overcoming Oscillations in Quantization-Aware Training
Markus Nagel
Marios Fournarakis
Yelysei Bondarenko
Tijmen Blankevoort
MQ
108
100
0
21 Mar 2022
Minimum Variance Unbiased N:M Sparsity for the Neural Gradients
Minimum Variance Unbiased N:M Sparsity for the Neural Gradients
Brian Chmiel
Itay Hubara
Ron Banner
Daniel Soudry
17
10
0
21 Mar 2022
QDrop: Randomly Dropping Quantization for Extremely Low-bit
  Post-Training Quantization
QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
Xiuying Wei
Ruihao Gong
Yuhang Li
Xianglong Liu
F. Yu
MQ
VLM
19
166
0
11 Mar 2022
Quantization in Layer's Input is Matter
Quantization in Layer's Input is Matter
Daning Cheng
Wenguang Chen
MQ
11
0
0
10 Feb 2022
Energy awareness in low precision neural networks
Energy awareness in low precision neural networks
Nurit Spingarn-Eliezer
Ron Banner
Elad Hoffer
Hilla Ben-Yaacov
T. Michaeli
38
0
0
06 Feb 2022
The Effect of Model Compression on Fairness in Facial Expression
  Recognition
The Effect of Model Compression on Fairness in Facial Expression Recognition
Samuil Stoychev
Hatice Gunes
CVBM
27
19
0
05 Jan 2022
Training Quantized Deep Neural Networks via Cooperative Coevolution
Training Quantized Deep Neural Networks via Cooperative Coevolution
Fu Peng
Shengcai Liu
Ning Lu
Ke Tang
MQ
18
1
0
23 Dec 2021
Accurate Neural Training with 4-bit Matrix Multiplications at Standard
  Formats
Accurate Neural Training with 4-bit Matrix Multiplications at Standard Formats
Brian Chmiel
Ron Banner
Elad Hoffer
Hilla Ben Yaacov
Daniel Soudry
MQ
25
22
0
19 Dec 2021
Towards Efficient Post-training Quantization of Pre-trained Language
  Models
Towards Efficient Post-training Quantization of Pre-trained Language Models
Haoli Bai
Lu Hou
Lifeng Shang
Xin Jiang
Irwin King
M. Lyu
MQ
73
47
0
30 Sep 2021
HPTQ: Hardware-Friendly Post Training Quantization
HPTQ: Hardware-Friendly Post Training Quantization
H. Habi
Reuven Peretz
Elad Cohen
Lior Dikstein
Oranit Dror
I. Diamant
Roy H. Jennings
Arnon Netzer
MQ
31
8
0
19 Sep 2021
Fine-grained Data Distribution Alignment for Post-Training Quantization
Fine-grained Data Distribution Alignment for Post-Training Quantization
Yunshan Zhong
Mingbao Lin
Mengzhao Chen
Ke Li
Yunhang Shen
Fei Chao
Yongjian Wu
Rongrong Ji
MQ
84
19
0
09 Sep 2021
Auto-Split: A General Framework of Collaborative Edge-Cloud AI
Auto-Split: A General Framework of Collaborative Edge-Cloud AI
Amin Banitalebi-Dehkordi
Naveen Vedula
J. Pei
Fei Xia
Lanjun Wang
Yong Zhang
22
89
0
30 Aug 2021
MOHAQ: Multi-Objective Hardware-Aware Quantization of Recurrent Neural
  Networks
MOHAQ: Multi-Objective Hardware-Aware Quantization of Recurrent Neural Networks
Nesma M. Rezk
Tomas Nordstrom
D. Stathis
Z. Ul-Abdin
E. Aksoy
A. Hemani
MQ
20
1
0
02 Aug 2021
Q-Rater: Non-Convex Optimization for Post-Training Uniform Quantization
Q-Rater: Non-Convex Optimization for Post-Training Uniform Quantization
Byeongwook Kim
Dongsoo Lee
Yeonju Ro
Yongkweon Jeon
S. Kwon
Baeseong Park
Daehwan Oh
MQ
10
1
0
05 May 2021
One Model for All Quantization: A Quantized Network Supporting Hot-Swap
  Bit-Width Adjustment
One Model for All Quantization: A Quantized Network Supporting Hot-Swap Bit-Width Adjustment
Qigong Sun
Xiufang Li
Yan Ren
Zhongjian Huang
Xu Liu
L. Jiao
Fang Liu
MQ
19
5
0
04 May 2021
Reliability-Aware Quantization for Anti-Aging NPUs
Reliability-Aware Quantization for Anti-Aging NPUs
Sami Salamin
Georgios Zervakis
Ourania Spantidi
Iraklis Anagnostopoulos
J. Henkel
H. Amrouch
6
13
0
08 Mar 2021
Accelerated Sparse Neural Training: A Provable and Efficient Method to
  Find N:M Transposable Masks
Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N:M Transposable Masks
Itay Hubara
Brian Chmiel
Moshe Island
Ron Banner
S. Naor
Daniel Soudry
50
110
0
16 Feb 2021
Confounding Tradeoffs for Neural Network Quantization
Confounding Tradeoffs for Neural Network Quantization
Sahaj Garg
Anirudh Jain
Joe Lou
Mitchell Nahmias
MQ
21
17
0
12 Feb 2021
BRECQ: Pushing the Limit of Post-Training Quantization by Block
  Reconstruction
BRECQ: Pushing the Limit of Post-Training Quantization by Block Reconstruction
Yuhang Li
Ruihao Gong
Xu Tan
Yang Yang
Peng Hu
Qi Zhang
F. Yu
Wei Wang
Shi Gu
MQ
19
414
0
10 Feb 2021
BinaryBERT: Pushing the Limit of BERT Quantization
BinaryBERT: Pushing the Limit of BERT Quantization
Haoli Bai
Wei Zhang
Lu Hou
Lifeng Shang
Jing Jin
Xin Jiang
Qun Liu
Michael Lyu
Irwin King
MQ
142
221
0
31 Dec 2020
Exploring Neural Networks Quantization via Layer-Wise Quantization
  Analysis
Exploring Neural Networks Quantization via Layer-Wise Quantization Analysis
Shachar Gluska
Mark Grobman
MQ
14
5
0
15 Dec 2020
A Tiny CNN Architecture for Medical Face Mask Detection for
  Resource-Constrained Endpoints
A Tiny CNN Architecture for Medical Face Mask Detection for Resource-Constrained Endpoints
P. Mohan
A. Paul
Abhay Chirania
CVBM
16
48
0
30 Nov 2020
AUSN: Approximately Uniform Quantization by Adaptively Superimposing
  Non-uniform Distribution for Deep Neural Networks
AUSN: Approximately Uniform Quantization by Adaptively Superimposing Non-uniform Distribution for Deep Neural Networks
Fangxin Liu
Wenbo Zhao
Yanzhi Wang
Changzhi Dai
Li Jiang
MQ
17
3
0
08 Jul 2020
Neural gradients are near-lognormal: improved quantized and sparse
  training
Neural gradients are near-lognormal: improved quantized and sparse training
Brian Chmiel
Liad Ben-Uri
Moran Shkolnik
Elad Hoffer
Ron Banner
Daniel Soudry
MQ
6
5
0
15 Jun 2020
Post-training Quantization with Multiple Points: Mixed Precision without
  Mixed Precision
Post-training Quantization with Multiple Points: Mixed Precision without Mixed Precision
Xingchao Liu
Mao Ye
Dengyong Zhou
Qiang Liu
MQ
8
42
0
20 Feb 2020
Robust Quantization: One Model to Rule Them All
Robust Quantization: One Model to Rule Them All
Moran Shkolnik
Brian Chmiel
Ron Banner
Gil Shomron
Yury Nahshan
A. Bronstein
U. Weiser
OOD
MQ
14
75
0
18 Feb 2020
Previous
12