ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1612.01064
  4. Cited By
Trained Ternary Quantization

Trained Ternary Quantization

4 December 2016
Chenzhuo Zhu
Song Han
Huizi Mao
W. Dally
    MQ
ArXivPDFHTML

Papers citing "Trained Ternary Quantization"

50 / 509 papers shown
Title
Neural Network Quantization for Efficient Inference: A Survey
Neural Network Quantization for Efficient Inference: A Survey
Olivia Weng
MQ
17
22
0
08 Dec 2021
Low-bit Quantization of Recurrent Neural Network Language Models Using
  Alternating Direction Methods of Multipliers
Low-bit Quantization of Recurrent Neural Network Language Models Using Alternating Direction Methods of Multipliers
Junhao Xu
Xie Chen
Shoukang Hu
Jianwei Yu
Xunying Liu
H. Meng
MQ
18
9
0
29 Nov 2021
AUTOKD: Automatic Knowledge Distillation Into A Student Architecture
  Family
AUTOKD: Automatic Knowledge Distillation Into A Student Architecture Family
Roy Henha Eyono
Fabio Maria Carlucci
P. Esperança
Binxin Ru
Phillip Torr
19
3
0
05 Nov 2021
Automatic Sleep Staging of EEG Signals: Recent Development, Challenges,
  and Future Directions
Automatic Sleep Staging of EEG Signals: Recent Development, Challenges, and Future Directions
Huy P Phan
Kaare B. Mikkelsen
11
93
0
03 Nov 2021
Arch-Net: Model Distillation for Architecture Agnostic Model Deployment
Arch-Net: Model Distillation for Architecture Agnostic Model Deployment
Weixin Xu
Zipeng Feng
Shuangkang Fang
Song Yuan
Yi Yang
Shuchang Zhou
MQ
16
1
0
01 Nov 2021
RMSMP: A Novel Deep Neural Network Quantization Framework with Row-wise
  Mixed Schemes and Multiple Precisions
RMSMP: A Novel Deep Neural Network Quantization Framework with Row-wise Mixed Schemes and Multiple Precisions
Sung-En Chang
Yanyu Li
Mengshu Sun
Weiwen Jiang
Sijia Liu
Yanzhi Wang
Xue Lin
MQ
8
10
0
30 Oct 2021
FAST: DNN Training Under Variable Precision Block Floating Point with
  Stochastic Rounding
FAST: DNN Training Under Variable Precision Block Floating Point with Stochastic Rounding
S. Zhang
Bradley McDanel
H. T. Kung
MQ
8
64
0
28 Oct 2021
MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning
MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning
Ji Lin
Wei-Ming Chen
Han Cai
Chuang Gan
Song Han
26
152
0
28 Oct 2021
MERCURY: Accelerating DNN Training By Exploiting Input Similarity
MERCURY: Accelerating DNN Training By Exploiting Input Similarity
Vahid Janfaza
Kevin Weston
Moein Razavi
Shantanu Mandal
Farabi Mahmud
Alex Hilty
A. Muzahid
39
5
0
28 Oct 2021
Demystifying and Generalizing BinaryConnect
Demystifying and Generalizing BinaryConnect
Abhishek Sharma
Yaoliang Yu
Eyyub Sari
Mahdi Zolnouri
V. Nia
MQ
17
8
0
25 Oct 2021
Applications and Techniques for Fast Machine Learning in Science
Applications and Techniques for Fast Machine Learning in Science
A. Deiana
Nhan Tran
Joshua C. Agar
Michaela Blott
G. D. Guglielmo
...
Ashish Sharma
S. Summers
Pietro Vischia
J. Vlimant
Olivia Weng
6
71
0
25 Oct 2021
When to Prune? A Policy towards Early Structural Pruning
When to Prune? A Policy towards Early Structural Pruning
Maying Shen
Pavlo Molchanov
Hongxu Yin
J. Álvarez
VLM
20
52
0
22 Oct 2021
HALP: Hardware-Aware Latency Pruning
HALP: Hardware-Aware Latency Pruning
Maying Shen
Hongxu Yin
Pavlo Molchanov
Lei Mao
Jianna Liu
J. Álvarez
VLM
30
13
0
20 Oct 2021
Sub-bit Neural Networks: Learning to Compress and Accelerate Binary
  Neural Networks
Sub-bit Neural Networks: Learning to Compress and Accelerate Binary Neural Networks
Yikai Wang
Yi Yang
Fuchun Sun
Anbang Yao
MQ
21
15
0
18 Oct 2021
Finding Everything within Random Binary Networks
Finding Everything within Random Binary Networks
Kartik K. Sreenivasan
Shashank Rajput
Jy-yong Sohn
Dimitris Papailiopoulos
14
10
0
18 Oct 2021
Network Augmentation for Tiny Deep Learning
Network Augmentation for Tiny Deep Learning
Han Cai
Chuang Gan
Ji Lin
Song Han
17
29
0
17 Oct 2021
BNAS v2: Learning Architectures for Binary Networks with Empirical
  Improvements
BNAS v2: Learning Architectures for Binary Networks with Empirical Improvements
Dahyun Kim
Kunal Pratap Singh
Jonghyun Choi
MQ
38
7
0
16 Oct 2021
Towards Mixed-Precision Quantization of Neural Networks via Constrained
  Optimization
Towards Mixed-Precision Quantization of Neural Networks via Constrained Optimization
Weihan Chen
Peisong Wang
Jian Cheng
MQ
33
61
0
13 Oct 2021
8-bit Optimizers via Block-wise Quantization
8-bit Optimizers via Block-wise Quantization
Tim Dettmers
M. Lewis
Sam Shleifer
Luke Zettlemoyer
MQ
17
268
0
06 Oct 2021
CBP: Backpropagation with constraint on weight precision using a
  pseudo-Lagrange multiplier method
CBP: Backpropagation with constraint on weight precision using a pseudo-Lagrange multiplier method
Guhyun Kim
D. Jeong
MQ
34
2
0
06 Oct 2021
Random matrices in service of ML footprint: ternary random features with
  no performance loss
Random matrices in service of ML footprint: ternary random features with no performance loss
Hafiz Tiomoko Ali
Zhenyu Liao
Romain Couillet
36
7
0
05 Oct 2021
Convolutional Neural Network Compression through Generalized Kronecker
  Product Decomposition
Convolutional Neural Network Compression through Generalized Kronecker Product Decomposition
Marawan Gamal Abdel Hameed
Marzieh S. Tahaei
A. Mosleh
V. Nia
39
25
0
29 Sep 2021
TSM: Temporal Shift Module for Efficient and Scalable Video
  Understanding on Edge Device
TSM: Temporal Shift Module for Efficient and Scalable Video Understanding on Edge Device
Ji Lin
Chuang Gan
Kuan-Chieh Jackson Wang
Song Han
38
64
0
27 Sep 2021
Distribution-sensitive Information Retention for Accurate Binary Neural
  Network
Distribution-sensitive Information Retention for Accurate Binary Neural Network
Haotong Qin
Xiangguo Zhang
Ruihao Gong
Yifu Ding
Yi Xu
Xianglong Liu
MQ
11
84
0
25 Sep 2021
ECQ$^{\text{x}}$: Explainability-Driven Quantization for Low-Bit and
  Sparse DNNs
ECQx^{\text{x}}x: Explainability-Driven Quantization for Low-Bit and Sparse DNNs
Daniel Becking
Maximilian Dreyer
Wojciech Samek
Karsten Müller
Sebastian Lapuschkin
MQ
93
13
0
09 Sep 2021
BioNetExplorer: Architecture-Space Exploration of Bio-Signal Processing
  Deep Neural Networks for Wearables
BioNetExplorer: Architecture-Space Exploration of Bio-Signal Processing Deep Neural Networks for Wearables
B. Prabakaran
Asima Akhtar
Semeen Rehman
Osman Hasan
Muhammad Shafique
11
9
0
07 Sep 2021
Efficient Visual Recognition with Deep Neural Networks: A Survey on
  Recent Advances and New Directions
Efficient Visual Recognition with Deep Neural Networks: A Survey on Recent Advances and New Directions
Yang Wu
Dingheng Wang
Xiaotong Lu
Fan Yang
Guoqi Li
W. Dong
Jianbo Shi
27
18
0
30 Aug 2021
Distance-aware Quantization
Distance-aware Quantization
Dohyung Kim
Junghyup Lee
Bumsub Ham
MQ
8
28
0
16 Aug 2021
Generalizable Mixed-Precision Quantization via Attribution Rank
  Preservation
Generalizable Mixed-Precision Quantization via Attribution Rank Preservation
Ziwei Wang
Han Xiao
Jiwen Lu
Jie Zhou
MQ
14
32
0
05 Aug 2021
Pruning Ternary Quantization
Danyang Liu
Xiangshan Chen
Jie Fu
Chen-li Ma
Xue Liu
MQ
31
0
0
23 Jul 2021
CREW: Computation Reuse and Efficient Weight Storage for
  Hardware-accelerated MLPs and RNNs
CREW: Computation Reuse and Efficient Weight Storage for Hardware-accelerated MLPs and RNNs
Marc Riera
J. Arnau
Antonio González
6
5
0
20 Jul 2021
A High-Performance Adaptive Quantization Approach for Edge CNN
  Applications
A High-Performance Adaptive Quantization Approach for Edge CNN Applications
Hsu-Hsun Chin
R. Tsay
Hsin-I Wu
MQ
6
5
0
18 Jul 2021
Training Compact CNNs for Image Classification using Dynamic-coded
  Filter Fusion
Training Compact CNNs for Image Classification using Dynamic-coded Filter Fusion
Mingbao Lin
Bohong Chen
Fei Chao
Rongrong Ji
VLM
25
23
0
14 Jul 2021
LANA: Latency Aware Network Acceleration
LANA: Latency Aware Network Acceleration
Pavlo Molchanov
Jimmy Hall
Hongxu Yin
Jan Kautz
Nicolò Fusi
Arash Vahdat
15
11
0
12 Jul 2021
Model compression as constrained optimization, with application to
  neural nets. Part V: combining compressions
Model compression as constrained optimization, with application to neural nets. Part V: combining compressions
Miguel Á. Carreira-Perpiñán
Yerlan Idelbayev
22
6
0
09 Jul 2021
$S^3$: Sign-Sparse-Shift Reparametrization for Effective Training of
  Low-bit Shift Networks
S3S^3S3: Sign-Sparse-Shift Reparametrization for Effective Training of Low-bit Shift Networks
Xinlin Li
Bang Liu
Yaoliang Yu
Wulong Liu
Chunjing Xu
V. Nia
MQ
28
5
0
07 Jul 2021
Content-Aware Convolutional Neural Networks
Content-Aware Convolutional Neural Networks
Yong Guo
Yaofo Chen
Mingkui Tan
K. Jia
Jian Chen
Jingdong Wang
30
8
0
30 Jun 2021
Post-Training Quantization for Vision Transformer
Post-Training Quantization for Vision Transformer
Zhenhua Liu
Yunhe Wang
Kai Han
Siwei Ma
Wen Gao
ViT
MQ
39
324
0
27 Jun 2021
Boggart: Towards General-Purpose Acceleration of Retrospective Video
  Analytics
Boggart: Towards General-Purpose Acceleration of Retrospective Video Analytics
Neil Agarwal
Ravi Netravali
21
14
0
21 Jun 2021
An Empirical Investigation into Deep and Shallow Rule Learning
An Empirical Investigation into Deep and Shallow Rule Learning
Florian Beck
Johannes Furnkranz
NAI
16
7
0
18 Jun 2021
Quantized Neural Networks via {-1, +1} Encoding Decomposition and
  Acceleration
Quantized Neural Networks via {-1, +1} Encoding Decomposition and Acceleration
Qigong Sun
Xiufang Li
Fanhua Shang
Hongying Liu
Kan Yang
L. Jiao
Zhouchen Lin
MQ
23
0
0
18 Jun 2021
Integer-Only Neural Network Quantization Scheme Based on
  Shift-Batch-Normalization
Integer-Only Neural Network Quantization Scheme Based on Shift-Batch-Normalization
Qingyu Guo
Yuan Wang
Xiaoxin Cui
MQ
11
2
0
28 May 2021
Is In-Domain Data Really Needed? A Pilot Study on Cross-Domain
  Calibration for Network Quantization
Is In-Domain Data Really Needed? A Pilot Study on Cross-Domain Calibration for Network Quantization
Haichao Yu
Linjie Yang
Humphrey Shi
OOD
MQ
18
5
0
16 May 2021
Model Pruning Based on Quantified Similarity of Feature Maps
Model Pruning Based on Quantified Similarity of Feature Maps
Zidu Wang
Xue-jun Liu
Long Huang
Yuxiang Chen
Yufei Zhang
Zhikang Lin
Rui Wang
18
16
0
13 May 2021
Binarized Weight Error Networks With a Transition Regularization Term
Binarized Weight Error Networks With a Transition Regularization Term
Savas Ozkan
G. Akar
MQ
21
1
0
09 May 2021
Modulating Regularization Frequency for Efficient Compression-Aware
  Model Training
Modulating Regularization Frequency for Efficient Compression-Aware Model Training
Dongsoo Lee
S. Kwon
Byeongwook Kim
Jeongin Yun
Baeseong Park
Yongkweon Jeon
11
0
0
05 May 2021
Q-Rater: Non-Convex Optimization for Post-Training Uniform Quantization
Q-Rater: Non-Convex Optimization for Post-Training Uniform Quantization
Byeongwook Kim
Dongsoo Lee
Yeonju Ro
Yongkweon Jeon
S. Kwon
Baeseong Park
Daehwan Oh
MQ
10
1
0
05 May 2021
Training Quantized Neural Networks to Global Optimality via Semidefinite
  Programming
Training Quantized Neural Networks to Global Optimality via Semidefinite Programming
Burak Bartan
Mert Pilanci
16
10
0
04 May 2021
Quantization of Deep Neural Networks for Accurate Edge Computing
Quantization of Deep Neural Networks for Accurate Edge Computing
Wentao Chen
Hailong Qiu
Zhuang Jian
Chutong Zhang
Yu Hu
Qing Lu
Tianchen Wang
Yiyu Shi
Meiping Huang
Xiaowe Xu
37
21
0
25 Apr 2021
Learning on Hardware: A Tutorial on Neural Network Accelerators and
  Co-Processors
Learning on Hardware: A Tutorial on Neural Network Accelerators and Co-Processors
Lukas Baischer
M. Wess
N. Taherinejad
12
12
0
19 Apr 2021
Previous
123456...91011
Next