ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1612.01064
  4. Cited By
Trained Ternary Quantization

Trained Ternary Quantization

4 December 2016
Chenzhuo Zhu
Song Han
Huizi Mao
W. Dally
    MQ
ArXivPDFHTML

Papers citing "Trained Ternary Quantization"

50 / 509 papers shown
Title
An FPGA Accelerated Method for Training Feed-forward Neural Networks
  Using Alternating Direction Method of Multipliers and LSMR
An FPGA Accelerated Method for Training Feed-forward Neural Networks Using Alternating Direction Method of Multipliers and LSMR
Seyedeh Niusha Alavi Foumani
Ce Guo
Wayne Luk
6
3
0
06 Sep 2020
Running Neural Networks on the NIC
Running Neural Networks on the NIC
G. Siracusano
Salvator Galea
D. Sanvito
Mohammad Malekzadeh
Hamed Haddadi
G. Antichi
R. Bifulco
11
25
0
04 Sep 2020
Transform Quantization for CNN (Convolutional Neural Network)
  Compression
Transform Quantization for CNN (Convolutional Neural Network) Compression
Sean I. Young
Wang Zhe
David S. Taubman
B. Girod
MQ
27
69
0
02 Sep 2020
GAN Slimming: All-in-One GAN Compression by A Unified Optimization
  Framework
GAN Slimming: All-in-One GAN Compression by A Unified Optimization Framework
Haotao Wang
Shupeng Gui
Haichuan Yang
Ji Liu
Zhangyang Wang
6
81
0
25 Aug 2020
One Weight Bitwidth to Rule Them All
One Weight Bitwidth to Rule Them All
Ting-Wu Chin
P. Chuang
Vikas Chandra
Diana Marculescu
MQ
17
25
0
22 Aug 2020
Channel-wise Hessian Aware trace-Weighted Quantization of Neural
  Networks
Channel-wise Hessian Aware trace-Weighted Quantization of Neural Networks
Xu Qian
Victor Li
Darren Crews
MQ
14
9
0
19 Aug 2020
Compression of Deep Learning Models for Text: A Survey
Compression of Deep Learning Models for Text: A Survey
Manish Gupta
Puneet Agrawal
VLM
MedIm
AI4CE
6
115
0
12 Aug 2020
FATNN: Fast and Accurate Ternary Neural Networks
FATNN: Fast and Accurate Ternary Neural Networks
Peng Chen
Bohan Zhuang
Chunhua Shen
MQ
4
15
0
12 Aug 2020
Hardware-Centric AutoML for Mixed-Precision Quantization
Hardware-Centric AutoML for Mixed-Precision Quantization
Kuan-Chieh Jackson Wang
Zhijian Liu
Yujun Lin
Ji Lin
Song Han
MQ
12
14
0
11 Aug 2020
PROFIT: A Novel Training Method for sub-4-bit MobileNet Models
PROFIT: A Novel Training Method for sub-4-bit MobileNet Models
Eunhyeok Park
S. Yoo
MQ
6
84
0
11 Aug 2020
Distance-Weighted Graph Neural Networks on FPGAs for Real-Time Particle
  Reconstruction in High Energy Physics
Distance-Weighted Graph Neural Networks on FPGAs for Real-Time Particle Reconstruction in High Energy Physics
Y. Iiyama
G. Cerminara
Abhijay Gupta
J. Kieseler
Vladimir Loncar
...
Miaoyuan Liu
K. Pedro
N. Tran
E. Kreinar
Zhenbin Wu
11
66
0
08 Aug 2020
Towards Lossless Binary Convolutional Neural Networks Using Piecewise
  Approximation
Towards Lossless Binary Convolutional Neural Networks Using Piecewise Approximation
Baozhou Zhu
Zaid Al-Ars
Wei Pan
MQ
14
8
0
08 Aug 2020
NASB: Neural Architecture Search for Binary Convolutional Neural
  Networks
NASB: Neural Architecture Search for Binary Convolutional Neural Networks
Baozhou Zhu
Zaid Al-Ars
P. Hofstee
MQ
15
23
0
08 Aug 2020
Fully Dynamic Inference with Deep Neural Networks
Fully Dynamic Inference with Deep Neural Networks
Wenhan Xia
Hongxu Yin
Xiaoliang Dai
N. Jha
3DH
BDL
26
38
0
29 Jul 2020
WrapNet: Neural Net Inference with Ultra-Low-Resolution Arithmetic
WrapNet: Neural Net Inference with Ultra-Low-Resolution Arithmetic
Renkun Ni
Hong-Min Chu
Oscar Castañeda
Ping Yeh-Chiang
Christoph Studer
Tom Goldstein
MQ
18
14
0
26 Jul 2020
Resource-Efficient Speech Mask Estimation for Multi-Channel Speech
  Enhancement
Resource-Efficient Speech Mask Estimation for Multi-Channel Speech Enhancement
Lukas Pfeifenberger
Matthias Zöhrer
Günther Schindler
Wolfgang Roth
Holger Fröning
Franz Pernkopf
9
1
0
22 Jul 2020
CovidDeep: SARS-CoV-2/COVID-19 Test Based on Wearable Medical Sensors
  and Efficient Neural Networks
CovidDeep: SARS-CoV-2/COVID-19 Test Based on Wearable Medical Sensors and Efficient Neural Networks
Shayan Hassantabar
Novati Stefano
Vishweshwar Ghanakota
A. Ferrari
G. Nicola
R. Bruno
I. Marino
Kenza Hamidouche
N. Jha
13
69
0
20 Jul 2020
MCUNet: Tiny Deep Learning on IoT Devices
MCUNet: Tiny Deep Learning on IoT Devices
Ji Lin
Wei-Ming Chen
Yujun Lin
J. Cohn
Chuang Gan
Song Han
56
471
0
20 Jul 2020
Search What You Want: Barrier Panelty NAS for Mixed Precision
  Quantization
Search What You Want: Barrier Panelty NAS for Mixed Precision Quantization
Haibao Yu
Qi Han
Jianbo Li
Jianping Shi
Guangliang Cheng
Bin Fan
MQ
16
61
0
20 Jul 2020
DBQ: A Differentiable Branch Quantizer for Lightweight Deep Neural
  Networks
DBQ: A Differentiable Branch Quantizer for Lightweight Deep Neural Networks
Hassan Dbouk
Hetul Sanghvi
M. Mehendale
Naresh R Shanbhag
MQ
14
9
0
19 Jul 2020
Channel-Level Variable Quantization Network for Deep Image Compression
Channel-Level Variable Quantization Network for Deep Image Compression
Zhisheng Zhong
Hiroaki Akutsu
Kiyoharu Aizawa
16
20
0
15 Jul 2020
T-Basis: a Compact Representation for Neural Networks
T-Basis: a Compact Representation for Neural Networks
Anton Obukhov
M. Rakhuba
Stamatios Georgoulis
Menelaos Kanakis
Dengxin Dai
Luc Van Gool
31
27
0
13 Jul 2020
Term Revealing: Furthering Quantization at Run Time on Quantized DNNs
Term Revealing: Furthering Quantization at Run Time on Quantized DNNs
H. T. Kung
Bradley McDanel
S. Zhang
MQ
13
9
0
13 Jul 2020
SGQuant: Squeezing the Last Bit on Graph Neural Networks with
  Specialized Quantization
SGQuant: Squeezing the Last Bit on Graph Neural Networks with Specialized Quantization
Boyuan Feng
Yuke Wang
Xu Li
Shu Yang
Xueqiao Peng
Yufei Ding
MQ
9
47
0
09 Jul 2020
Operation-Aware Soft Channel Pruning using Differentiable Masks
Operation-Aware Soft Channel Pruning using Differentiable Masks
Minsoo Kang
Bohyung Han
AAML
25
138
0
08 Jul 2020
EasyQuant: Post-training Quantization via Scale Optimization
EasyQuant: Post-training Quantization via Scale Optimization
Di Wu
Qingming Tang
Yongle Zhao
Ming Zhang
Ying Fu
Debing Zhang
MQ
17
75
0
30 Jun 2020
Efficient Integer-Arithmetic-Only Convolutional Neural Networks
Efficient Integer-Arithmetic-Only Convolutional Neural Networks
Hengrui Zhao
Dong Liu
Houqiang Li
MQ
12
4
0
21 Jun 2020
Faster Secure Data Mining via Distributed Homomorphic Encryption
Faster Secure Data Mining via Distributed Homomorphic Encryption
Junyi Li
Heng-Chiao Huang
FedML
18
20
0
17 Jun 2020
APQ: Joint Search for Network Architecture, Pruning and Quantization
  Policy
APQ: Joint Search for Network Architecture, Pruning and Quantization Policy
Tianzhe Wang
Kuan-Chieh Jackson Wang
Han Cai
Ji Lin
Zhijian Liu
Song Han
MQ
28
174
0
15 Jun 2020
Optimal Lottery Tickets via SubsetSum: Logarithmic Over-Parameterization
  is Sufficient
Optimal Lottery Tickets via SubsetSum: Logarithmic Over-Parameterization is Sufficient
Ankit Pensia
Shashank Rajput
Alliot Nagle
Harit Vishwakarma
Dimitris Papailiopoulos
17
102
0
14 Jun 2020
An Overview of Neural Network Compression
An Overview of Neural Network Compression
James OÑeill
AI4CE
45
98
0
05 Jun 2020
FBNetV3: Joint Architecture-Recipe Search using Predictor Pretraining
FBNetV3: Joint Architecture-Recipe Search using Predictor Pretraining
Xiaoliang Dai
Alvin Wan
Peizhao Zhang
Bichen Wu
Zijian He
...
Kan Chen
Yuandong Tian
Matthew Yu
Peter Vajda
Joseph E. Gonzalez
10
72
0
03 Jun 2020
Quantized Neural Networks: Characterization and Holistic Optimization
Quantized Neural Networks: Characterization and Holistic Optimization
Yoonho Boo
Sungho Shin
Wonyong Sung
MQ
40
8
0
31 May 2020
PruneNet: Channel Pruning via Global Importance
PruneNet: Channel Pruning via Global Importance
A. Khetan
Zohar S. Karnin
10
11
0
22 May 2020
Cross-filter compression for CNN inference acceleration
Cross-filter compression for CNN inference acceleration
Fuyuan Lyu
Shien Zhu
Weichen Liu
MQ
12
0
0
18 May 2020
VecQ: Minimal Loss DNN Model Compression With Vectorized Weight
  Quantization
VecQ: Minimal Loss DNN Model Compression With Vectorized Weight Quantization
Cheng Gong
Yao Chen
Ye Lu
Tao Li
Cong Hao
Deming Chen
MQ
6
44
0
18 May 2020
A flexible, extensible software framework for model compression based on
  the LC algorithm
A flexible, extensible software framework for model compression based on the LC algorithm
Yerlan Idelbayev
Miguel Á. Carreira-Perpiñán
4
9
0
15 May 2020
Binarizing MobileNet via Evolution-based Searching
Binarizing MobileNet via Evolution-based Searching
Hai T. Phan
Zechun Liu
Dang T. Huynh
Marios Savvides
Kwang-Ting Cheng
Zhiqiang Shen
3DV
MQ
24
43
0
13 May 2020
schuBERT: Optimizing Elements of BERT
schuBERT: Optimizing Elements of BERT
A. Khetan
Zohar S. Karnin
23
30
0
09 May 2020
Data-Free Network Quantization With Adversarial Knowledge Distillation
Data-Free Network Quantization With Adversarial Knowledge Distillation
Yoojin Choi
Jihwan P. Choi
Mostafa El-Khamy
Jungwon Lee
MQ
6
119
0
08 May 2020
Constructing Accurate and Efficient Deep Spiking Neural Networks with
  Double-threshold and Augmented Schemes
Constructing Accurate and Efficient Deep Spiking Neural Networks with Double-threshold and Augmented Schemes
Qiang Yu
Chenxiang Ma
Shiming Song
Gaoyan Zhang
J. Dang
Kay Chen Tan
12
49
0
05 May 2020
Learning for Microrobot Exploration: Model-based Locomotion,
  Sparse-robust Navigation, and Low-power Deep Classification
Learning for Microrobot Exploration: Model-based Locomotion, Sparse-robust Navigation, and Low-power Deep Classification
Nathan Lambert
Farhan Toddywala
B. Liao
Eric Zhu
Lydia Lee
K. Pister
11
0
0
27 Apr 2020
Fully Embedding Fast Convolutional Networks on Pixel Processor Arrays
Fully Embedding Fast Convolutional Networks on Pixel Processor Arrays
Laurie Bose
Jianing Chen
S. Carey
Piotr Dudek
W. Mayol-Cuevas
12
37
0
27 Apr 2020
Lite Transformer with Long-Short Range Attention
Lite Transformer with Long-Short Range Attention
Zhanghao Wu
Zhijian Liu
Ji Lin
Yujun Lin
Song Han
16
317
0
24 Apr 2020
DyNet: Dynamic Convolution for Accelerating Convolutional Neural
  Networks
DyNet: Dynamic Convolution for Accelerating Convolutional Neural Networks
Yikang Zhang
Jian Andrew Zhang
Qiang-qiang Wang
Zhaobai Zhong
6
87
0
22 Apr 2020
Integer Quantization for Deep Learning Inference: Principles and
  Empirical Evaluation
Integer Quantization for Deep Learning Inference: Principles and Empirical Evaluation
Hao Wu
Patrick Judd
Xiaojie Zhang
Mikhail Isaev
Paulius Micikevicius
MQ
24
340
0
20 Apr 2020
MuBiNN: Multi-Level Binarized Recurrent Neural Network for EEG signal
  Classification
MuBiNN: Multi-Level Binarized Recurrent Neural Network for EEG signal Classification
Seyed Ahmad Mirsalari
Sima Sinaei
M. Salehi
Masoud Daneshtalab
MQ
6
5
0
19 Apr 2020
Efficient Synthesis of Compact Deep Neural Networks
Efficient Synthesis of Compact Deep Neural Networks
Wenhan Xia
Hongxu Yin
N. Jha
16
3
0
18 Apr 2020
Entropy-Based Modeling for Estimating Soft Errors Impact on Binarized
  Neural Network Inference
Entropy-Based Modeling for Estimating Soft Errors Impact on Binarized Neural Network Inference
N. Khoshavi
S. Sargolzaei
A. Roohi
Connor Broyles
Yu Bi
AAML
12
1
0
10 Apr 2020
A Learning Framework for n-bit Quantized Neural Networks toward FPGAs
A Learning Framework for n-bit Quantized Neural Networks toward FPGAs
Jun Chen
L. Liu
Yong Liu
Xianfang Zeng
MQ
10
26
0
06 Apr 2020
Previous
123456...91011
Next