ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1810.01018
  4. Cited By
Simultaneously Optimizing Weight and Quantizer of Ternary Neural Network
  using Truncated Gaussian Approximation

Simultaneously Optimizing Weight and Quantizer of Ternary Neural Network using Truncated Gaussian Approximation

2 October 2018
Zhezhi He
Deliang Fan
    MQ
ArXivPDFHTML

Papers citing "Simultaneously Optimizing Weight and Quantizer of Ternary Neural Network using Truncated Gaussian Approximation"

29 / 29 papers shown
Title
Ternarization of Vision Language Models for use on edge devices
Ternarization of Vision Language Models for use on edge devices
Ben Crulis
Cyril de Runz
Barthélémy Serres
Gilles Venturini
VLM
55
0
0
07 Apr 2025
Quasar-ViT: Hardware-Oriented Quantization-Aware Architecture Search for
  Vision Transformers
Quasar-ViT: Hardware-Oriented Quantization-Aware Architecture Search for Vision Transformers
Zhengang Li
Alec Lu
Yanyue Xie
Zhenglun Kong
Mengshu Sun
...
Peiyan Dong
Caiwen Ding
Yanzhi Wang
Xue Lin
Zhenman Fang
32
5
0
25 Jul 2024
SpikeZIP-TF: Conversion is All You Need for Transformer-based SNN
SpikeZIP-TF: Conversion is All You Need for Transformer-based SNN
Kang You
Zekai Xu
Chen Nie
Zhijie Deng
Qinghai Guo
Xiang Wang
Zhezhi He
38
10
0
05 Jun 2024
SupeRBNN: Randomized Binary Neural Network Using Adiabatic
  Superconductor Josephson Devices
SupeRBNN: Randomized Binary Neural Network Using Adiabatic Superconductor Josephson Devices
Z. Li
Geng Yuan
Tomoharu Yamauchi
Zabihi Masoud
Yanyue Xie
...
Xulong Tang
Nobuyuki Yoshikawa
Devesh Tiwari
Yanzhi Wang
O. Chen
MQ
6
4
0
21 Sep 2023
MinUn: Accurate ML Inference on Microcontrollers
MinUn: Accurate ML Inference on Microcontrollers
Shikhar Jaiswal
R. Goli
Aayan Kumar
Vivek Seshadri
Rahul Sharma
21
2
0
29 Oct 2022
Auto-ViT-Acc: An FPGA-Aware Automatic Acceleration Framework for Vision
  Transformer with Mixed-Scheme Quantization
Auto-ViT-Acc: An FPGA-Aware Automatic Acceleration Framework for Vision Transformer with Mixed-Scheme Quantization
Z. Li
Mengshu Sun
Alec Lu
Haoyu Ma
Geng Yuan
...
Yanyu Li
M. Leeser
Zhangyang Wang
Xue Lin
Zhenman Fang
ViT
MQ
14
49
0
10 Aug 2022
Quantum Neural Network Compression
Quantum Neural Network Compression
Zhirui Hu
Peiyan Dong
Zhepeng Wang
Youzuo Lin
Yanzhi Wang
Weiwen Jiang
GNN
25
28
0
04 Jul 2022
Rethinking Spatial Invariance of Convolutional Networks for Object
  Counting
Rethinking Spatial Invariance of Convolutional Networks for Object Counting
Zhi-Qi Cheng
Qi Dai
Hong Li
JingKuan Song
Xiao-Jun Wu
Alexander G. Hauptmann
3DPC
31
95
0
10 Jun 2022
BiBERT: Accurate Fully Binarized BERT
BiBERT: Accurate Fully Binarized BERT
Haotong Qin
Yifu Ding
Mingyuan Zhang
Qing Yan
Aishan Liu
Qingqing Dang
Ziwei Liu
Xianglong Liu
MQ
14
93
0
12 Mar 2022
N3H-Core: Neuron-designed Neural Network Accelerator via FPGA-based
  Heterogeneous Computing Cores
N3H-Core: Neuron-designed Neural Network Accelerator via FPGA-based Heterogeneous Computing Cores
Yu Gong
Zhihang Xu
Zhezhi He
Weifeng Zhang
Xiaobing Tu
Xiaoyao Liang
Li Jiang
17
13
0
15 Dec 2021
RMSMP: A Novel Deep Neural Network Quantization Framework with Row-wise
  Mixed Schemes and Multiple Precisions
RMSMP: A Novel Deep Neural Network Quantization Framework with Row-wise Mixed Schemes and Multiple Precisions
Sung-En Chang
Yanyu Li
Mengshu Sun
Weiwen Jiang
Sijia Liu
Yanzhi Wang
Xue Lin
MQ
8
10
0
30 Oct 2021
Distribution-sensitive Information Retention for Accurate Binary Neural
  Network
Distribution-sensitive Information Retention for Accurate Binary Neural Network
Haotong Qin
Xiangguo Zhang
Ruihao Gong
Yifu Ding
Yi Xu
Xianglong Liu
MQ
14
84
0
25 Sep 2021
Quantization and Deployment of Deep Neural Networks on Microcontrollers
Quantization and Deployment of Deep Neural Networks on Microcontrollers
Pierre-Emmanuel Novac
G. B. Hacene
Alain Pegatoquet
Benoit Miramond
Vincent Gripon
MQ
20
116
0
27 May 2021
Learning on Hardware: A Tutorial on Neural Network Accelerators and
  Co-Processors
Learning on Hardware: A Tutorial on Neural Network Accelerators and Co-Processors
Lukas Baischer
M. Wess
N. Taherinejad
14
12
0
19 Apr 2021
ReCU: Reviving the Dead Weights in Binary Neural Networks
ReCU: Reviving the Dead Weights in Binary Neural Networks
Zihan Xu
Mingbao Lin
Jianzhuang Liu
Jie Chen
Ling Shao
Yue Gao
Yonghong Tian
Rongrong Ji
MQ
19
81
0
23 Mar 2021
Distribution Adaptive INT8 Quantization for Training CNNs
Distribution Adaptive INT8 Quantization for Training CNNs
Kang Zhao
Sida Huang
Pan Pan
Yinghan Li
Yingya Zhang
Zhenyu Gu
Yinghui Xu
MQ
14
63
0
09 Feb 2021
Continual Learning of Generative Models with Limited Data: From
  Wasserstein-1 Barycenter to Adaptive Coalescence
Continual Learning of Generative Models with Limited Data: From Wasserstein-1 Barycenter to Adaptive Coalescence
M. Dedeoglu
Sen Lin
Zhaofeng Zhang
Junshan Zhang
9
1
0
22 Jan 2021
Mix and Match: A Novel FPGA-Centric Deep Neural Network Quantization
  Framework
Mix and Match: A Novel FPGA-Centric Deep Neural Network Quantization Framework
Sung-En Chang
Yanyu Li
Mengshu Sun
Runbin Shi
Hayden Kwok-Hay So
Xuehai Qian
Yanzhi Wang
Xue Lin
MQ
18
82
0
08 Dec 2020
MetaGater: Fast Learning of Conditional Channel Gated Networks via
  Federated Meta-Learning
MetaGater: Fast Learning of Conditional Channel Gated Networks via Federated Meta-Learning
Sen Lin
Li Yang
Zhezhi He
Deliang Fan
Junshan Zhang
FedML
AI4CE
9
5
0
25 Nov 2020
MSP: An FPGA-Specific Mixed-Scheme, Multi-Precision Deep Neural Network
  Quantization Framework
MSP: An FPGA-Specific Mixed-Scheme, Multi-Precision Deep Neural Network Quantization Framework
Sung-En Chang
Yanyu Li
Mengshu Sun
Weiwen Jiang
Runbin Shi
Xue Lin
Yanzhi Wang
MQ
16
7
0
16 Sep 2020
SoFAr: Shortcut-based Fractal Architectures for Binary Convolutional
  Neural Networks
SoFAr: Shortcut-based Fractal Architectures for Binary Convolutional Neural Networks
Baozhou Zhu
P. Hofstee
Jinho Lee
Zaid Al-Ars
MQ
6
2
0
11 Sep 2020
Towards Lossless Binary Convolutional Neural Networks Using Piecewise
  Approximation
Towards Lossless Binary Convolutional Neural Networks Using Piecewise Approximation
Baozhou Zhu
Zaid Al-Ars
Wei Pan
MQ
14
8
0
08 Aug 2020
SparseTrain: Exploiting Dataflow Sparsity for Efficient Convolutional
  Neural Networks Training
SparseTrain: Exploiting Dataflow Sparsity for Efficient Convolutional Neural Networks Training
Pengcheng Dai
Jianlei Yang
Xucheng Ye
Xingzhou Cheng
Junyu Luo
Linghao Song
Yiran Chen
Weisheng Zhao
17
21
0
21 Jul 2020
Learning Sparse & Ternary Neural Networks with Entropy-Constrained
  Trained Ternarization (EC2T)
Learning Sparse & Ternary Neural Networks with Entropy-Constrained Trained Ternarization (EC2T)
Arturo Marbán
Daniel Becking
Simon Wiedemann
Wojciech Samek
MQ
6
11
0
02 Apr 2020
Filter Sketch for Network Pruning
Filter Sketch for Network Pruning
Mingbao Lin
Liujuan Cao
Shaojie Li
QiXiang Ye
Yonghong Tian
Jianzhuang Liu
Q. Tian
Rongrong Ji
CLIP
3DPC
12
82
0
23 Jan 2020
Towards Unified INT8 Training for Convolutional Neural Network
Towards Unified INT8 Training for Convolutional Neural Network
Feng Zhu
Ruihao Gong
F. Yu
Xianglong Liu
Yanfei Wang
Zhelong Li
Xiuqi Yang
Junjie Yan
MQ
27
151
0
29 Dec 2019
Forward and Backward Information Retention for Accurate Binary Neural
  Networks
Forward and Backward Information Retention for Accurate Binary Neural Networks
Haotong Qin
Ruihao Gong
Xianglong Liu
Mingzhu Shen
Ziran Wei
F. Yu
Jingkuan Song
MQ
117
324
0
24 Sep 2019
Non-Structured DNN Weight Pruning -- Is It Beneficial in Any Platform?
Non-Structured DNN Weight Pruning -- Is It Beneficial in Any Platform?
Xiaolong Ma
Sheng Lin
Shaokai Ye
Zhezhi He
Linfeng Zhang
...
Deliang Fan
Xuehai Qian
X. Lin
Kaisheng Ma
Yanzhi Wang
MQ
13
92
0
03 Jul 2019
Weight Normalization based Quantization for Deep Neural Network
  Compression
Weight Normalization based Quantization for Deep Neural Network Compression
Wenhong Cai
Wu-Jun Li
16
14
0
01 Jul 2019
1