ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1612.01064
  4. Cited By
Trained Ternary Quantization

Trained Ternary Quantization

4 December 2016
Chenzhuo Zhu
Song Han
Huizi Mao
W. Dally
    MQ
ArXivPDFHTML

Papers citing "Trained Ternary Quantization"

50 / 509 papers shown
Title
Towards Efficient Graph Convolutional Networks for Point Cloud Handling
Towards Efficient Graph Convolutional Networks for Point Cloud Handling
Yawei Li
He Chen
Zhaopeng Cui
Radu Timofte
Marc Pollefeys
Gregory S. Chirikjian
Luc Van Gool
3DPC
GNN
12
24
0
12 Apr 2021
NullaNet Tiny: Ultra-low-latency DNN Inference Through Fixed-function
  Combinational Logic
NullaNet Tiny: Ultra-low-latency DNN Inference Through Fixed-function Combinational Logic
M. Nazemi
A. Fayyazi
Amirhossein Esmaili
Atharva Khare
Soheil Nazar Shahsavani
Massoud Pedram
17
13
0
07 Apr 2021
Network Quantization with Element-wise Gradient Scaling
Network Quantization with Element-wise Gradient Scaling
Junghyup Lee
Dohyung Kim
Bumsub Ham
MQ
8
115
0
02 Apr 2021
Bit-Mixer: Mixed-precision networks with runtime bit-width selection
Bit-Mixer: Mixed-precision networks with runtime bit-width selection
Adrian Bulat
Georgios Tzimiropoulos
MQ
13
27
0
31 Mar 2021
Compressing 1D Time-Channel Separable Convolutions using Sparse Random
  Ternary Matrices
Compressing 1D Time-Channel Separable Convolutions using Sparse Random Ternary Matrices
Gonçalo Mordido
Matthijs Van Keirsbilck
A. Keller
30
6
0
31 Mar 2021
DNN Quantization with Attention
DNN Quantization with Attention
G. B. Hacene
Lukas Mauch
Stefan Uhlich
Fabien Cardinaux
MQ
6
2
0
24 Mar 2021
Pufferfish: Communication-efficient Models At No Extra Cost
Pufferfish: Communication-efficient Models At No Extra Cost
Hongyi Wang
Saurabh Agarwal
Dimitris Papailiopoulos
11
56
0
05 Mar 2021
Artificial Neural Networks generated by Low Discrepancy Sequences
Artificial Neural Networks generated by Low Discrepancy Sequences
A. Keller
Matthijs Van Keirsbilck
17
5
0
05 Mar 2021
Anycost GANs for Interactive Image Synthesis and Editing
Anycost GANs for Interactive Image Synthesis and Editing
Ji Lin
Richard Y. Zhang
F. Ganz
Song Han
Jun-Yan Zhu
36
83
0
04 Mar 2021
QuPeL: Quantized Personalization with Applications to Federated Learning
QuPeL: Quantized Personalization with Applications to Federated Learning
Kaan Ozkara
Navjot Singh
Deepesh Data
Suhas Diggavi
FedML
22
5
0
23 Feb 2021
VS-Quant: Per-vector Scaled Quantization for Accurate Low-Precision
  Neural Network Inference
VS-Quant: Per-vector Scaled Quantization for Accurate Low-Precision Neural Network Inference
Steve Dai
Rangharajan Venkatesan
Haoxing Ren
B. Zimmer
W. Dally
Brucek Khailany
MQ
25
67
0
08 Feb 2021
Rethinking Floating Point Overheads for Mixed Precision DNN Accelerators
Rethinking Floating Point Overheads for Mixed Precision DNN Accelerators
Hamzah Abdel-Aziz
Ali Shafiee
J. Shin
A. Pedram
Joseph Hassoun
MQ
28
10
0
27 Jan 2021
Pruning and Quantization for Deep Neural Network Acceleration: A Survey
Pruning and Quantization for Deep Neural Network Acceleration: A Survey
Tailin Liang
C. Glossner
Lei Wang
Shaobo Shi
Xiaotong Zhang
MQ
124
671
0
24 Jan 2021
Generative Zero-shot Network Quantization
Generative Zero-shot Network Quantization
Xiangyu He
Qinghao Hu
Peisong Wang
Jian Cheng
GAN
MQ
23
23
0
21 Jan 2021
Accelerating Deep Learning Inference via Learned Caches
Accelerating Deep Learning Inference via Learned Caches
Arjun Balasubramanian
Adarsh Kumar
Yuhan Liu
Han Cao
Shivaram Venkataraman
Aditya Akella
22
17
0
18 Jan 2021
Network Automatic Pruning: Start NAP and Take a Nap
Network Automatic Pruning: Start NAP and Take a Nap
Wenyuan Zeng
Yuwen Xiong
R. Urtasun
26
9
0
17 Jan 2021
Direct Quantization for Training Highly Accurate Low Bit-width Deep
  Neural Networks
Direct Quantization for Training Highly Accurate Low Bit-width Deep Neural Networks
Ziquan Liu
Wuguannan Yao
Qiao Li
Antoni B. Chan
MQ
14
9
0
26 Dec 2020
Hybrid and Non-Uniform quantization methods using retro synthesis data
  for efficient inference
Hybrid and Non-Uniform quantization methods using retro synthesis data for efficient inference
Gvsl Tej Pratap
R. Kumar
MQ
16
1
0
26 Dec 2020
Adaptive Precision Training for Resource Constrained Devices
Adaptive Precision Training for Resource Constrained Devices
Tian Huang
Tao Luo
Joey Tianyi Zhou
26
5
0
23 Dec 2020
Robustness and Transferability of Universal Attacks on Compressed Models
Robustness and Transferability of Universal Attacks on Compressed Models
Alberto G. Matachana
Kenneth T. Co
Luis Muñoz-González
David Martínez
Emil C. Lupu
AAML
16
10
0
10 Dec 2020
Recurrence of Optimum for Training Weight and Activation Quantized
  Networks
Recurrence of Optimum for Training Weight and Activation Quantized Networks
Ziang Long
Penghang Yin
Jack Xin
MQ
27
3
0
10 Dec 2020
Mix and Match: A Novel FPGA-Centric Deep Neural Network Quantization
  Framework
Mix and Match: A Novel FPGA-Centric Deep Neural Network Quantization Framework
Sung-En Chang
Yanyu Li
Mengshu Sun
Runbin Shi
Hayden Kwok-Hay So
Xuehai Qian
Yanzhi Wang
Xue Lin
MQ
18
82
0
08 Dec 2020
Going Beyond Classification Accuracy Metrics in Model Compression
Going Beyond Classification Accuracy Metrics in Model Compression
Vinu Joseph
Shoaib Ahmed Siddiqui
Aditya Bhaskara
Ganesh Gopalakrishnan
Saurav Muralidharan
M. Garland
Sheraz Ahmed
Andreas Dengel
37
17
0
03 Dec 2020
Bringing AI To Edge: From Deep Learning's Perspective
Bringing AI To Edge: From Deep Learning's Perspective
Di Liu
Hao Kong
Xiangzhong Luo
Weichen Liu
Ravi Subramaniam
42
116
0
25 Nov 2020
Distributed Additive Encryption and Quantization for Privacy Preserving
  Federated Deep Learning
Distributed Additive Encryption and Quantization for Privacy Preserving Federated Deep Learning
Hangyu Zhu
Rui Wang
Yaochu Jin
K. Liang
Jianting Ning
FedML
14
46
0
25 Nov 2020
Learning Quantized Neural Nets by Coarse Gradient Method for Non-linear
  Classification
Learning Quantized Neural Nets by Coarse Gradient Method for Non-linear Classification
Ziang Long
Penghang Yin
Jack Xin
MQ
22
3
0
23 Nov 2020
MixMix: All You Need for Data-Free Compression Are Feature and Data
  Mixing
MixMix: All You Need for Data-Free Compression Are Feature and Data Mixing
Yuhang Li
Feng Zhu
Ruihao Gong
Mingzhu Shen
Xin Dong
F. Yu
Shaoqing Lu
Shi Gu
MQ
23
38
0
19 Nov 2020
Larq Compute Engine: Design, Benchmark, and Deploy State-of-the-Art
  Binarized Neural Networks
Larq Compute Engine: Design, Benchmark, and Deploy State-of-the-Art Binarized Neural Networks
T. Bannink
Arash Bakhtiari
Adam Hillier
Lukas Geiger
T. D. Bruin
Leon Overweel
J. Neeven
K. Helwegen
3DV
MQ
13
36
0
18 Nov 2020
Analytical aspects of non-differentiable neural networks
Analytical aspects of non-differentiable neural networks
G. P. Leonardi
Matteo Spallanzani
9
1
0
03 Nov 2020
Sparsity-Control Ternary Weight Networks
Sparsity-Control Ternary Weight Networks
Xiang Deng
Zhongfei Zhang
MQ
10
8
0
01 Nov 2020
Permute, Quantize, and Fine-tune: Efficient Compression of Neural
  Networks
Permute, Quantize, and Fine-tune: Efficient Compression of Neural Networks
Julieta Martinez
Jashan Shewakramani
Ting Liu
Ioan Andrei Bârsan
Wenyuan Zeng
R. Urtasun
MQ
13
30
0
29 Oct 2020
MARS: Multi-macro Architecture SRAM CIM-Based Accelerator with
  Co-designed Compressed Neural Networks
MARS: Multi-macro Architecture SRAM CIM-Based Accelerator with Co-designed Compressed Neural Networks
Syuan-Hao Sie
Jye-Luen Lee
Yi-Ren Chen
Chih-Cheng Lu
C. Hsieh
Meng-Fan Chang
K. Tang
17
14
0
24 Oct 2020
Robustness-aware 2-bit quantization with real-time performance for
  neural network
Robustness-aware 2-bit quantization with real-time performance for neural network
Xiaobin Li
Hongxu Jiang
Shuangxi Huang
Fangzheng Tian
MQ
11
1
0
19 Oct 2020
FTBNN: Rethinking Non-linearity for 1-bit CNNs and Going Beyond
FTBNN: Rethinking Non-linearity for 1-bit CNNs and Going Beyond
Z. Su
Linpu Fang
Deke Guo
Duwen Hu
M. Pietikäinen
Li Liu
MQ
13
3
0
19 Oct 2020
An Investigation on Different Underlying Quantization Schemes for
  Pre-trained Language Models
An Investigation on Different Underlying Quantization Schemes for Pre-trained Language Models
Zihan Zhao
Yuncong Liu
Lu Chen
Qi Liu
Rao Ma
Kai Yu
MQ
8
12
0
14 Oct 2020
A Very Compact Embedded CNN Processor Design Based on Logarithmic
  Computing
A Very Compact Embedded CNN Processor Design Based on Logarithmic Computing
Tsung-Ying Lu
Hsu-Hsun Chin
Hsin-I Wu
R. Tsay
MQ
14
1
0
13 Oct 2020
TUTOR: Training Neural Networks Using Decision Rules as Model Priors
TUTOR: Training Neural Networks Using Decision Rules as Model Priors
Shayan Hassantabar
Prerit Terway
N. Jha
22
10
0
12 Oct 2020
High-Capacity Expert Binary Networks
High-Capacity Expert Binary Networks
Adrian Bulat
Brais Martínez
Georgios Tzimiropoulos
MQ
15
57
0
07 Oct 2020
Improving Network Slimming with Nonconvex Regularization
Improving Network Slimming with Nonconvex Regularization
Kevin Bui
Fredrick Park
Shuai Zhang
Y. Qi
Jack Xin
16
9
0
03 Oct 2020
Binary Neural Networks for Memory-Efficient and Effective Visual Place
  Recognition in Changing Environments
Binary Neural Networks for Memory-Efficient and Effective Visual Place Recognition in Changing Environments
Bruno Ferrarini
Michael Milford
Klaus D. McDonald-Maier
Shoaib Ehsan
MQ
22
22
0
01 Oct 2020
Stochastic Precision Ensemble: Self-Knowledge Distillation for Quantized
  Deep Neural Networks
Stochastic Precision Ensemble: Self-Knowledge Distillation for Quantized Deep Neural Networks
Yoonho Boo
Sungho Shin
Jungwook Choi
Wonyong Sung
MQ
14
29
0
30 Sep 2020
Self-grouping Convolutional Neural Networks
Self-grouping Convolutional Neural Networks
Qingbei Guo
Xiaojun Wu
J. Kittler
Zhiquan Feng
19
22
0
29 Sep 2020
NITI: Training Integer Neural Networks Using Integer-only Arithmetic
NITI: Training Integer Neural Networks Using Integer-only Arithmetic
Maolin Wang
Seyedramin Rasoulinezhad
Philip H. W. Leong
Hayden Kwok-Hay So
MQ
10
39
0
28 Sep 2020
Learned Low Precision Graph Neural Networks
Learned Low Precision Graph Neural Networks
Yiren Zhao
Duo Wang
Daniel Bates
Robert D. Mullins
M. Jamnik
Pietro Lió
GNN
31
34
0
19 Sep 2020
Searching for Low-Bit Weights in Quantized Neural Networks
Searching for Low-Bit Weights in Quantized Neural Networks
Zhaohui Yang
Yunhe Wang
Kai Han
Chunjing Xu
Chao Xu
Dacheng Tao
Chang Xu
MQ
14
82
0
18 Sep 2020
MEAL V2: Boosting Vanilla ResNet-50 to 80%+ Top-1 Accuracy on ImageNet
  without Tricks
MEAL V2: Boosting Vanilla ResNet-50 to 80%+ Top-1 Accuracy on ImageNet without Tricks
Zhiqiang Shen
Marios Savvides
17
63
0
17 Sep 2020
MSP: An FPGA-Specific Mixed-Scheme, Multi-Precision Deep Neural Network
  Quantization Framework
MSP: An FPGA-Specific Mixed-Scheme, Multi-Precision Deep Neural Network Quantization Framework
Sung-En Chang
Yanyu Li
Mengshu Sun
Weiwen Jiang
Runbin Shi
Xue Lin
Yanzhi Wang
MQ
13
7
0
16 Sep 2020
QuantNet: Learning to Quantize by Learning within Fully Differentiable
  Framework
QuantNet: Learning to Quantize by Learning within Fully Differentiable Framework
Junjie Liu
Dongchao Wen
Deyu Wang
Wei Tao
Tse-Wei Chen
Kinya Osa
Masami Kato
MQ
12
3
0
10 Sep 2020
FleXOR: Trainable Fractional Quantization
FleXOR: Trainable Fractional Quantization
Dongsoo Lee
S. Kwon
Byeongwook Kim
Yongkweon Jeon
Baeseong Park
Jeongin Yun
MQ
9
13
0
09 Sep 2020
Binarized Neural Architecture Search for Efficient Object Recognition
Binarized Neural Architecture Search for Efficient Object Recognition
Hanlin Chen
Lian Zhuo
Baochang Zhang
Xiawu Zheng
Jianzhuang Liu
Rongrong Ji
David Doermann
G. Guo
MQ
6
18
0
08 Sep 2020
Previous
12345...91011
Next