ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1811.08886
  4. Cited By
HAQ: Hardware-Aware Automated Quantization with Mixed Precision
v1v2v3 (latest)

HAQ: Hardware-Aware Automated Quantization with Mixed Precision

Computer Vision and Pattern Recognition (CVPR), 2018
21 November 2018
Kuan-Chieh Wang
Zhijian Liu
Chengyue Wu
Ji Lin
Song Han
    MQ
ArXiv (abs)PDFHTML

Papers citing "HAQ: Hardware-Aware Automated Quantization with Mixed Precision"

50 / 462 papers shown
Title
Automated Model Compression by Jointly Applied Pruning and Quantization
Automated Model Compression by Jointly Applied Pruning and Quantization
Wenting Tang
Xingxing Wei
Yue Liu
MQ
91
7
0
12 Nov 2020
Resource-Aware Pareto-Optimal Automated Machine Learning Platform
Resource-Aware Pareto-Optimal Automated Machine Learning Platform
Yao Yang
Andrew Nam
M. Nasr-Azadani
Teresa Tung
114
7
0
30 Oct 2020
Permute, Quantize, and Fine-tune: Efficient Compression of Neural
  Networks
Permute, Quantize, and Fine-tune: Efficient Compression of Neural NetworksComputer Vision and Pattern Recognition (CVPR), 2020
Julieta Martinez
Jashan Shewakramani
Ting Liu
Ioan Andrei Bârsan
Wenyuan Zeng
R. Urtasun
MQ
264
33
0
29 Oct 2020
An Investigation on Different Underlying Quantization Schemes for
  Pre-trained Language Models
An Investigation on Different Underlying Quantization Schemes for Pre-trained Language Models
Zihan Zhao
Yuncong Liu
Lu Chen
Qi Liu
Rao Ma
Kai Yu
MQ
98
13
0
14 Oct 2020
Once Quantization-Aware Training: High Performance Extremely Low-bit
  Architecture Search
Once Quantization-Aware Training: High Performance Extremely Low-bit Architecture SearchIEEE International Conference on Computer Vision (ICCV), 2020
Mingzhu Shen
Feng Liang
Yazhe Niu
Yuhang Li
Chuming Li
Chen Lin
F. Yu
Junjie Yan
Wanli Ouyang
MQ
180
45
0
09 Oct 2020
Online Knowledge Distillation via Multi-branch Diversity Enhancement
Online Knowledge Distillation via Multi-branch Diversity EnhancementAsian Conference on Computer Vision (ACCV), 2020
Zheng Li
Ying Huang
Defang Chen
Tianren Luo
Ning Cai
Zhigeng Pan
198
34
0
02 Oct 2020
MSP: An FPGA-Specific Mixed-Scheme, Multi-Precision Deep Neural Network
  Quantization Framework
MSP: An FPGA-Specific Mixed-Scheme, Multi-Precision Deep Neural Network Quantization Framework
Sung-En Chang
Yanyu Li
Mengshu Sun
Weiwen Jiang
Runbin Shi
Xue Lin
Yanzhi Wang
MQ
122
8
0
16 Sep 2020
FleXOR: Trainable Fractional Quantization
FleXOR: Trainable Fractional QuantizationNeural Information Processing Systems (NeurIPS), 2020
Dongsoo Lee
S. Kwon
Byeongwook Kim
Yongkweon Jeon
Baeseong Park
Jeongin Yun
MQ
193
13
0
09 Sep 2020
Layer-specific Optimization for Mixed Data Flow with Mixed Precision in
  FPGA Design for CNN-based Object Detectors
Layer-specific Optimization for Mixed Data Flow with Mixed Precision in FPGA Design for CNN-based Object Detectors
Duy-Thanh Nguyen
Hyun Kim
Hyuk-Jae Lee
MQ
95
76
0
03 Sep 2020
Transform Quantization for CNN (Convolutional Neural Network)
  Compression
Transform Quantization for CNN (Convolutional Neural Network) CompressionIEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2020
Sean I. Young
Wang Zhe
David S. Taubman
B. Girod
MQ
250
88
0
02 Sep 2020
GAN Slimming: All-in-One GAN Compression by A Unified Optimization
  Framework
GAN Slimming: All-in-One GAN Compression by A Unified Optimization FrameworkEuropean Conference on Computer Vision (ECCV), 2020
Haotao Wang
Shupeng Gui
Haichuan Yang
Ji Liu
Zinan Lin
228
85
0
25 Aug 2020
Matching Guided Distillation
Matching Guided Distillation
Kaiyu Yue
Jiangfan Deng
Feng Zhou
159
57
0
23 Aug 2020
One Weight Bitwidth to Rule Them All
One Weight Bitwidth to Rule Them All
Ting-Wu Chin
P. Chuang
Vikas Chandra
Diana Marculescu
MQ
169
26
0
22 Aug 2020
Channel-wise Hessian Aware trace-Weighted Quantization of Neural
  Networks
Channel-wise Hessian Aware trace-Weighted Quantization of Neural Networks
Xu Qian
Victor Li
Darren Crews
MQ
100
9
0
19 Aug 2020
Leveraging Automated Mixed-Low-Precision Quantization for tiny edge
  microcontrollers
Leveraging Automated Mixed-Low-Precision Quantization for tiny edge microcontrollers
Manuele Rusci
Marco Fariselli
Alessandro Capotondi
Luca Benini
MQ
105
24
0
12 Aug 2020
Degree-Quant: Quantization-Aware Training for Graph Neural Networks
Degree-Quant: Quantization-Aware Training for Graph Neural NetworksInternational Conference on Learning Representations (ICLR), 2020
Shyam A. Tailor
Javier Fernandez-Marques
Nicholas D. Lane
GNNMQ
257
166
0
11 Aug 2020
HAPI: Hardware-Aware Progressive Inference
HAPI: Hardware-Aware Progressive Inference
Stefanos Laskaridis
Stylianos I. Venieris
Hyeji Kim
Nicholas D. Lane
164
49
0
10 Aug 2020
Modeling Data Reuse in Deep Neural Networks by Taking Data-Types into
  Cognizance
Modeling Data Reuse in Deep Neural Networks by Taking Data-Types into CognizanceIEEE transactions on computers (IEEE Trans. Comput.), 2020
N. Jha
Sparsh Mittal
122
16
0
06 Aug 2020
Continuous-in-Depth Neural Networks
Continuous-in-Depth Neural Networks
A. Queiruga
N. Benjamin Erichson
D. Taylor
Michael W. Mahoney
249
54
0
05 Aug 2020
Searching Efficient 3D Architectures with Sparse Point-Voxel Convolution
Searching Efficient 3D Architectures with Sparse Point-Voxel ConvolutionEuropean Conference on Computer Vision (ECCV), 2020
Haotian Tang
Zhijian Liu
Shengyu Zhao
Chengyue Wu
Ji Lin
Hanrui Wang
Song Han
3DPC
411
754
0
31 Jul 2020
WrapNet: Neural Net Inference with Ultra-Low-Resolution Arithmetic
WrapNet: Neural Net Inference with Ultra-Low-Resolution Arithmetic
Renkun Ni
Hong-Min Chu
Oscar Castañeda
Ping Yeh-Chiang
Christoph Studer
Tom Goldstein
MQ
122
15
0
26 Jul 2020
Differentiable Joint Pruning and Quantization for Hardware Efficiency
Differentiable Joint Pruning and Quantization for Hardware Efficiency
Ying Wang
Yadong Lu
Tijmen Blankevoort
MQ
249
82
0
20 Jul 2020
MCUNet: Tiny Deep Learning on IoT Devices
MCUNet: Tiny Deep Learning on IoT Devices
Ji Lin
Wei-Ming Chen
Chengyue Wu
J. Cohn
Chuang Gan
Song Han
344
582
0
20 Jul 2020
Search What You Want: Barrier Panelty NAS for Mixed Precision
  Quantization
Search What You Want: Barrier Panelty NAS for Mixed Precision Quantization
Haibao Yu
Qi Han
Jianbo Li
Jianping Shi
Guangliang Cheng
Bin Fan
MQ
139
66
0
20 Jul 2020
HMQ: Hardware Friendly Mixed Precision Quantization Block for CNNs
HMQ: Hardware Friendly Mixed Precision Quantization Block for CNNs
H. Habi
Roy H. Jennings
Arnon Netzer
MQ
177
74
0
20 Jul 2020
DBQ: A Differentiable Branch Quantizer for Lightweight Deep Neural
  Networks
DBQ: A Differentiable Branch Quantizer for Lightweight Deep Neural NetworksEuropean Conference on Computer Vision (ECCV), 2020
Hassan Dbouk
Hetul Sanghvi
M. Mehendale
Naresh R Shanbhag
MQ
119
9
0
19 Jul 2020
Standing on the Shoulders of Giants: Hardware and Neural Architecture
  Co-Search with Hot Start
Standing on the Shoulders of Giants: Hardware and Neural Architecture Co-Search with Hot StartIEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems (TCAD), 2020
Weiwen Jiang
Lei Yang
Sakyasingha Dasgupta
Jiaxi Hu
Yiyu Shi
166
66
0
17 Jul 2020
Optimizing Memory Placement using Evolutionary Graph Reinforcement
  Learning
Optimizing Memory Placement using Evolutionary Graph Reinforcement LearningInternational Conference on Learning Representations (ICLR), 2020
Shauharda Khadka
Estelle Aflalo
Mattias Marder
Avrech Ben-David
Santiago Miret
Shie Mannor
Tamir Hazan
Hanlin Tang
Somdeb Majumdar
GNN
150
13
0
14 Jul 2020
AUSN: Approximately Uniform Quantization by Adaptively Superimposing
  Non-uniform Distribution for Deep Neural Networks
AUSN: Approximately Uniform Quantization by Adaptively Superimposing Non-uniform Distribution for Deep Neural Networks
Fangxin Liu
Wenbo Zhao
Yanzhi Wang
Changzhi Dai
Li Jiang
MQ
122
4
0
08 Jul 2020
FracBits: Mixed Precision Quantization via Fractional Bit-Widths
FracBits: Mixed Precision Quantization via Fractional Bit-Widths
Linjie Yang
Qing Jin
MQ
194
91
0
04 Jul 2020
Bit Error Robustness for Energy-Efficient DNN Accelerators
Bit Error Robustness for Energy-Efficient DNN Accelerators
David Stutz
Nandhini Chandramoorthy
Matthias Hein
Bernt Schiele
MQ
221
1
0
24 Jun 2020
Automatic heterogeneous quantization of deep neural networks for
  low-latency inference on the edge for particle detectors
Automatic heterogeneous quantization of deep neural networks for low-latency inference on the edge for particle detectors
C. Coelho
Aki Kuusela
Shane Li
Zhuang Hao
T. Aarrestad
Vladimir Loncar
J. Ngadiuba
M. Pierini
Adrian Alan Pol
S. Summers
MQ
230
209
0
15 Jun 2020
Automated Design Space Exploration for optimised Deployment of DNN on
  Arm Cortex-A CPUs
Automated Design Space Exploration for optimised Deployment of DNN on Arm Cortex-A CPUs
Miguel de Prado
Andrew Mundy
Rabia Saeed
Maurizo Denna
Nuria Pazos
Luca Benini
158
11
0
09 Jun 2020
EDCompress: Energy-Aware Model Compression for Dataflows
EDCompress: Energy-Aware Model Compression for Dataflows
Zhehui Wang
Yaoyu Zhang
Qiufeng Wang
Rick Siow Mong Goh
131
2
0
08 Jun 2020
Novel Adaptive Binary Search Strategy-First Hybrid Pyramid- and
  Clustering-Based CNN Filter Pruning Method without Parameters Setting
Novel Adaptive Binary Search Strategy-First Hybrid Pyramid- and Clustering-Based CNN Filter Pruning Method without Parameters Setting
K. Chung
Yu-Lun Chang
Bo-Wei Tsai
74
0
0
08 Jun 2020
Conditional Neural Architecture Search
Conditional Neural Architecture Search
Sheng-Chun Kao
Arun Ramamurthy
Reed Williams
T. Krishna
95
0
0
06 Jun 2020
Generative Design of Hardware-aware DNNs
Generative Design of Hardware-aware DNNs
Sheng-Chun Kao
Arun Ramamurthy
T. Krishna
MQ
101
2
0
06 Jun 2020
Machine Learning Systems for Intelligent Services in the IoT: A Survey
Wiebke Toussaint
Aaron Yi Ding
LRM
169
0
0
29 May 2020
A Feature-map Discriminant Perspective for Pruning Deep Neural Networks
A Feature-map Discriminant Perspective for Pruning Deep Neural Networks
Zejiang Hou
S. Kung
81
5
0
28 May 2020
Accelerating Neural Network Inference by Overflow Aware Quantization
Accelerating Neural Network Inference by Overflow Aware Quantization
Hongwei Xie
Shuo Zhang
Huanghao Ding
Yafei Song
Baitao Shao
Conggang Hu
Lingyi Cai
Mingyang Li
MQ
59
0
0
27 May 2020
VecQ: Minimal Loss DNN Model Compression With Vectorized Weight
  Quantization
VecQ: Minimal Loss DNN Model Compression With Vectorized Weight Quantization
Cheng Gong
Yao Chen
Ye Lu
Tao Li
Cong Hao
Deming Chen
MQ
150
58
0
18 May 2020
Bayesian Bits: Unifying Quantization and Pruning
Bayesian Bits: Unifying Quantization and Pruning
M. V. Baalen
Christos Louizos
Markus Nagel
Rana Ali Amjad
Ying Wang
Tijmen Blankevoort
Max Welling
MQ
318
130
0
14 May 2020
Data-Free Network Quantization With Adversarial Knowledge Distillation
Data-Free Network Quantization With Adversarial Knowledge Distillation
Yoojin Choi
Jihwan P. Choi
Mostafa El-Khamy
Jungwon Lee
MQ
170
136
0
08 May 2020
Lite Transformer with Long-Short Range Attention
Lite Transformer with Long-Short Range AttentionInternational Conference on Learning Representations (ICLR), 2020
Zhanghao Wu
Zhijian Liu
Ji Lin
Chengyue Wu
Song Han
156
356
0
24 Apr 2020
Automatic low-bit hybrid quantization of neural networks through meta
  learning
Automatic low-bit hybrid quantization of neural networks through meta learning
Tao Wang
Junsong Wang
Chang Xu
Chao Xue
MQ
64
2
0
24 Apr 2020
Intermittent Inference with Nonuniformly Compressed Multi-Exit Neural
  Network for Energy Harvesting Powered Devices
Intermittent Inference with Nonuniformly Compressed Multi-Exit Neural Network for Energy Harvesting Powered Devices
Yawen Wu
Zhepeng Wang
Zhenge Jia
Yiyu Shi
Jiaxi Hu
146
59
0
23 Apr 2020
LSQ+: Improving low-bit quantization through learnable offsets and
  better initialization
LSQ+: Improving low-bit quantization through learnable offsets and better initialization
Brandon Smart
Jinwon Lee
Markus Nagel
Tijmen Blankevoort
Nojun Kwak
MQ
168
272
0
20 Apr 2020
Learning Low-rank Deep Neural Networks via Singular Vector Orthogonality
  Regularization and Singular Value Sparsification
Learning Low-rank Deep Neural Networks via Singular Vector Orthogonality Regularization and Singular Value Sparsification
Huanrui Yang
Minxue Tang
W. Wen
Feng Yan
Daniel Hu
Ang Li
Xue Yang
Yiran Chen
130
78
0
20 Apr 2020
Efficient Synthesis of Compact Deep Neural Networks
Efficient Synthesis of Compact Deep Neural NetworksDesign Automation Conference (DAC), 2020
Wenhan Xia
Hongxu Yin
N. Jha
122
3
0
18 Apr 2020
Rethinking Differentiable Search for Mixed-Precision Neural Networks
Rethinking Differentiable Search for Mixed-Precision Neural NetworksComputer Vision and Pattern Recognition (CVPR), 2020
Zhaowei Cai
Nuno Vasconcelos
MQ
136
138
0
13 Apr 2020
Previous
123...10789
Next