ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1612.01064
  4. Cited By
Trained Ternary Quantization

Trained Ternary Quantization

4 December 2016
Chenzhuo Zhu
Song Han
Huizi Mao
W. Dally
    MQ
ArXivPDFHTML

Papers citing "Trained Ternary Quantization"

50 / 509 papers shown
Title
Learning Sparse & Ternary Neural Networks with Entropy-Constrained
  Trained Ternarization (EC2T)
Learning Sparse & Ternary Neural Networks with Entropy-Constrained Trained Ternarization (EC2T)
Arturo Marbán
Daniel Becking
Simon Wiedemann
Wojciech Samek
MQ
6
11
0
02 Apr 2020
A Survey of Convolutional Neural Networks: Analysis, Applications, and
  Prospects
A Survey of Convolutional Neural Networks: Analysis, Applications, and Prospects
Zewen Li
Wenjie Yang
Shouheng Peng
Fan Liu
HAI
3DV
54
2,595
0
01 Apr 2020
How Not to Give a FLOP: Combining Regularization and Pruning for
  Efficient Inference
How Not to Give a FLOP: Combining Regularization and Pruning for Efficient Inference
Tai Vu
Emily Wen
Roy Nehoran
7
5
0
30 Mar 2020
Training Binary Neural Networks with Real-to-Binary Convolutions
Training Binary Neural Networks with Real-to-Binary Convolutions
Brais Martínez
Jing Yang
Adrian Bulat
Georgios Tzimiropoulos
MQ
9
226
0
25 Mar 2020
GAN Compression: Efficient Architectures for Interactive Conditional
  GANs
GAN Compression: Efficient Architectures for Interactive Conditional GANs
Muyang Li
Ji Lin
Yaoyao Ding
Zhijian Liu
Jun-Yan Zhu
Song Han
GAN
15
2
0
19 Mar 2020
Efficient Bitwidth Search for Practical Mixed Precision Neural Network
Efficient Bitwidth Search for Practical Mixed Precision Neural Network
Yuhang Li
Wei Wang
Haoli Bai
Ruihao Gong
Xin Dong
F. Yu
MQ
8
20
0
17 Mar 2020
A flexible framework for communication-efficient machine learning: from
  HPC to IoT
A flexible framework for communication-efficient machine learning: from HPC to IoT
Sarit Khirirat
Sindri Magnússon
Arda Aytekin
M. Johansson
14
7
0
13 Mar 2020
Ternary Compression for Communication-Efficient Federated Learning
Ternary Compression for Communication-Efficient Federated Learning
Jinjin Xu
W. Du
Ran Cheng
Wangli He
Yaochu Jin
MQ
FedML
34
174
0
07 Mar 2020
ReActNet: Towards Precise Binary Neural Network with Generalized
  Activation Functions
ReActNet: Towards Precise Binary Neural Network with Generalized Activation Functions
Zechun Liu
Zhiqiang Shen
Marios Savvides
Kwang-Ting Cheng
MQ
14
347
0
07 Mar 2020
Propagating Asymptotic-Estimated Gradients for Low Bitwidth Quantized
  Neural Networks
Propagating Asymptotic-Estimated Gradients for Low Bitwidth Quantized Neural Networks
Jun Chen
Yong Liu
Hao Zhang
Shengnan Hou
Jian Yang
MQ
11
7
0
04 Mar 2020
BATS: Binary ArchitecTure Search
BATS: Binary ArchitecTure Search
Adrian Bulat
Brais Martínez
Georgios Tzimiropoulos
MQ
9
67
0
03 Mar 2020
WaveQ: Gradient-Based Deep Quantization of Neural Networks through
  Sinusoidal Adaptive Regularization
WaveQ: Gradient-Based Deep Quantization of Neural Networks through Sinusoidal Adaptive Regularization
Ahmed T. Elthakeb
Prannoy Pilligundla
Fatemehsadat Mireshghallah
T. Elgindi
Charles-Alban Deledalle
H. Esmaeilzadeh
MQ
17
10
0
29 Feb 2020
Quantized Neural Network Inference with Precision Batching
Quantized Neural Network Inference with Precision Batching
Maximilian Lam
Zachary Yedidia
Colby R. Banbury
Vijay Janapa Reddi
MQ
20
1
0
26 Feb 2020
Post-training Quantization with Multiple Points: Mixed Precision without
  Mixed Precision
Post-training Quantization with Multiple Points: Mixed Precision without Mixed Precision
Xingchao Liu
Mao Ye
Dengyong Zhou
Qiang Liu
MQ
8
42
0
20 Feb 2020
SYMOG: learning symmetric mixture of Gaussian modes for improved
  fixed-point quantization
SYMOG: learning symmetric mixture of Gaussian modes for improved fixed-point quantization
Lukas Enderich
Fabian Timm
Wolfram Burgard
MQ
6
6
0
19 Feb 2020
Precision Gating: Improving Neural Network Efficiency with Dynamic
  Dual-Precision Activations
Precision Gating: Improving Neural Network Efficiency with Dynamic Dual-Precision Activations
Yichi Zhang
Ritchie Zhao
Weizhe Hua
N. Xu
G. E. Suh
Zhiru Zhang
MQ
82
27
0
17 Feb 2020
Learning Architectures for Binary Networks
Learning Architectures for Binary Networks
Dahyun Kim
Kunal Pratap Singh
Jonghyun Choi
MQ
22
44
0
17 Feb 2020
Switchable Precision Neural Networks
Switchable Precision Neural Networks
Luis Guerra
Bohan Zhuang
Ian Reid
Tom Drummond
MQ
12
20
0
07 Feb 2020
Exponential discretization of weights of neural network connections in
  pre-trained neural networks
Exponential discretization of weights of neural network connections in pre-trained neural networks
M. Malsagov
E. Khayrov
M. Pushkareva
I. Karandashev
9
5
0
03 Feb 2020
Widening and Squeezing: Towards Accurate and Efficient QNNs
Widening and Squeezing: Towards Accurate and Efficient QNNs
Chuanjian Liu
Kai Han
Yunhe Wang
Hanting Chen
Qi Tian
Chunjing Xu
MQ
6
0
0
03 Feb 2020
SQWA: Stochastic Quantized Weight Averaging for Improving the
  Generalization Capability of Low-Precision Deep Neural Networks
SQWA: Stochastic Quantized Weight Averaging for Improving the Generalization Capability of Low-Precision Deep Neural Networks
Sungho Shin
Yoonho Boo
Wonyong Sung
MQ
15
3
0
02 Feb 2020
Post-Training Piecewise Linear Quantization for Deep Neural Networks
Post-Training Piecewise Linear Quantization for Deep Neural Networks
Jun Fang
Ali Shafiee
Hamzah Abdel-Aziz
D. Thorsley
Georgios Georgiadis
Joseph Hassoun
MQ
10
143
0
31 Jan 2020
Variational Dropout Sparsification for Particle Identification speed-up
Variational Dropout Sparsification for Particle Identification speed-up
Artem Sergeevich Ryzhikov
D. Derkach
M. Hushchyn
12
0
0
21 Jan 2020
MeliusNet: Can Binary Neural Networks Achieve MobileNet-level Accuracy?
MeliusNet: Can Binary Neural Networks Achieve MobileNet-level Accuracy?
Joseph Bethge
Christian Bartz
Haojin Yang
Ying Chen
Christoph Meinel
MQ
20
91
0
16 Jan 2020
Least squares binary quantization of neural networks
Least squares binary quantization of neural networks
Hadi Pouransari
Zhucheng Tu
Oncel Tuzel
MQ
12
32
0
09 Jan 2020
Resource-Efficient Neural Networks for Embedded Systems
Resource-Efficient Neural Networks for Embedded Systems
Wolfgang Roth
Günther Schindler
Lukas Pfeifenberger
Robert Peharz
Sebastian Tschiatschek
Holger Fröning
Franz Pernkopf
Zoubin Ghahramani
26
47
0
07 Jan 2020
Sparse Weight Activation Training
Sparse Weight Activation Training
Md Aamir Raihan
Tor M. Aamodt
32
72
0
07 Jan 2020
RPR: Random Partition Relaxation for Training; Binary and Ternary Weight
  Neural Networks
RPR: Random Partition Relaxation for Training; Binary and Ternary Weight Neural Networks
Lukas Cavigelli
Luca Benini
MQ
13
9
0
04 Jan 2020
Fractional Skipping: Towards Finer-Grained Dynamic CNN Inference
Fractional Skipping: Towards Finer-Grained Dynamic CNN Inference
Jianghao Shen
Y. Fu
Yue Wang
Pengfei Xu
Zhangyang Wang
Yingyan Lin
MQ
14
95
0
03 Jan 2020
ZeroQ: A Novel Zero Shot Quantization Framework
ZeroQ: A Novel Zero Shot Quantization Framework
Yaohui Cai
Z. Yao
Zhen Dong
A. Gholami
Michael W. Mahoney
Kurt Keutzer
MQ
30
389
0
01 Jan 2020
Towards Efficient Training for Neural Network Quantization
Towards Efficient Training for Neural Network Quantization
Qing Jin
Linjie Yang
Zhenyu A. Liao
MQ
11
42
0
21 Dec 2019
AdaBits: Neural Network Quantization with Adaptive Bit-Widths
AdaBits: Neural Network Quantization with Adaptive Bit-Widths
Qing Jin
Linjie Yang
Zhenyu A. Liao
MQ
16
123
0
20 Dec 2019
FQ-Conv: Fully Quantized Convolution for Efficient and Accurate
  Inference
FQ-Conv: Fully Quantized Convolution for Efficient and Accurate Inference
Bram-Ernst Verhoef
Nathan Laubeuf
S. Cosemans
P. Debacker
Ioannis A. Papistas
A. Mallik
D. Verkest
MQ
11
16
0
19 Dec 2019
Dreaming to Distill: Data-free Knowledge Transfer via DeepInversion
Dreaming to Distill: Data-free Knowledge Transfer via DeepInversion
Hongxu Yin
Pavlo Molchanov
Zhizhong Li
J. Álvarez
Arun Mallya
Derek Hoiem
N. Jha
Jan Kautz
15
551
0
18 Dec 2019
STEERAGE: Synthesis of Neural Networks Using Architecture Search and
  Grow-and-Prune Methods
STEERAGE: Synthesis of Neural Networks Using Architecture Search and Grow-and-Prune Methods
Shayan Hassantabar
Xiaoliang Dai
N. Jha
3DV
17
17
0
12 Dec 2019
Dynamic Convolution: Attention over Convolution Kernels
Dynamic Convolution: Attention over Convolution Kernels
Yinpeng Chen
Xiyang Dai
Mengchen Liu
Dongdong Chen
Lu Yuan
Zicheng Liu
11
867
0
07 Dec 2019
Sampling-Free Learning of Bayesian Quantized Neural Networks
Sampling-Free Learning of Bayesian Quantized Neural Networks
Jiahao Su
Milan Cvitkovic
Furong Huang
BDL
MQ
UQCV
8
7
0
06 Dec 2019
RTN: Reparameterized Ternary Network
RTN: Reparameterized Ternary Network
Yuhang Li
Xin Dong
S. Zhang
Haoli Bai
Yuanpeng Chen
Wei Wang
MQ
11
28
0
04 Dec 2019
QKD: Quantization-aware Knowledge Distillation
QKD: Quantization-aware Knowledge Distillation
Jangho Kim
Yash Bhalgat
Jinwon Lee
Chirag I. Patel
Nojun Kwak
MQ
16
63
0
28 Nov 2019
Quantization Networks
Quantization Networks
Jiwei Yang
Xu Shen
Jun Xing
Xinmei Tian
Houqiang Li
Bing Deng
Jianqiang Huang
Xiansheng Hua
MQ
25
338
0
21 Nov 2019
Distributed Low Precision Training Without Mixed Precision
Distributed Low Precision Training Without Mixed Precision
Zehua Cheng
Weiyan Wang
Yan Pan
Thomas Lukasiewicz
MQ
16
5
0
18 Nov 2019
Real-time ultra-low power ECG anomaly detection using an event-driven
  neuromorphic processor
Real-time ultra-low power ECG anomaly detection using an event-driven neuromorphic processor
F. Bauer
Dylan R. Muir
Giacomo Indiveri
13
95
0
13 Nov 2019
A Programmable Approach to Neural Network Compression
A Programmable Approach to Neural Network Compression
Vinu Joseph
Saurav Muralidharan
Animesh Garg
M. Garland
Ganesh Gopalakrishnan
6
10
0
06 Nov 2019
Ternary MobileNets via Per-Layer Hybrid Filter Banks
Ternary MobileNets via Per-Layer Hybrid Filter Banks
Dibakar Gope
Jesse G. Beu
Urmish Thakker
Matthew Mattina
MQ
22
15
0
04 Nov 2019
LUTNet: Learning FPGA Configurations for Highly Efficient Neural Network
  Inference
LUTNet: Learning FPGA Configurations for Highly Efficient Neural Network Inference
Erwei Wang
James J. Davis
P. Cheung
G. Constantinides
MQ
9
41
0
24 Oct 2019
Automatic Generation of Multi-precision Multi-arithmetic CNN
  Accelerators for FPGAs
Automatic Generation of Multi-precision Multi-arithmetic CNN Accelerators for FPGAs
Yiren Zhao
Xitong Gao
Xuan Guo
Junyi Liu
Erwei Wang
Robert D. Mullins
P. Cheung
G. Constantinides
Chengzhong Xu
MQ
17
31
0
21 Oct 2019
Automatic Neural Network Compression by Sparsity-Quantization Joint
  Learning: A Constrained Optimization-based Approach
Automatic Neural Network Compression by Sparsity-Quantization Joint Learning: A Constrained Optimization-based Approach
Haichuan Yang
Shupeng Gui
Yuhao Zhu
Ji Liu
MQ
12
5
0
14 Oct 2019
EDEN: Enabling Energy-Efficient, High-Performance Deep Neural Network
  Inference Using Approximate DRAM
EDEN: Enabling Energy-Efficient, High-Performance Deep Neural Network Inference Using Approximate DRAM
Skanda Koppula
Lois Orosa
A. G. Yaglikçi
Roknoddin Azizi
Taha Shahroodi
Konstantinos Kanellopoulos
O. Mutlu
17
105
0
12 Oct 2019
DiabDeep: Pervasive Diabetes Diagnosis based on Wearable Medical Sensors
  and Efficient Neural Networks
DiabDeep: Pervasive Diabetes Diagnosis based on Wearable Medical Sensors and Efficient Neural Networks
Hongxu Yin
Bilal Mukadam
Xiaoliang Dai
N. Jha
20
47
0
11 Oct 2019
Structured Pruning of Large Language Models
Structured Pruning of Large Language Models
Ziheng Wang
Jeremy Wohlwend
Tao Lei
24
280
0
10 Oct 2019
Previous
123...567...91011
Next