Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1605.04711
Cited By
Ternary Weight Networks
16 May 2016
Fengfu Li
Bin Liu
Xiaoxing Wang
Bo-Wen Zhang
Junchi Yan
MQ
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Ternary Weight Networks"
50 / 63 papers shown
Title
Optimizing LLMs for Resource-Constrained Environments: A Survey of Model Compression Techniques
Sanjay Surendranath Girija
Shashank Kapoor
Lakshit Arora
Dipen Pradhan
Aman Raj
Ankit Shetgaonkar
52
0
0
05 May 2025
BackSlash: Rate Constrained Optimized Training of Large Language Models
Jun Wu
Jiangtao Wen
Yuxing Han
34
0
0
23 Apr 2025
Online Difficulty Filtering for Reasoning Oriented Reinforcement Learning
Sanghwan Bae
Jiwoo Hong
Min Young Lee
Hanbyul Kim
Jeongyeon Nam
Donghyun Kwak
OffRL
LRM
48
0
0
04 Apr 2025
Forget the Data and Fine-Tuning! Just Fold the Network to Compress
Dong Wang
Haris Šikić
Lothar Thiele
O. Saukh
44
0
0
17 Feb 2025
BILLNET: A Binarized Conv3D-LSTM Network with Logic-gated residual architecture for hardware-efficient video inference
Van Thien Nguyen
William Guicquero
Gilles Sicard
3DV
MQ
74
2
0
24 Jan 2025
MOGNET: A Mux-residual quantized Network leveraging Online-Generated weights
Van Thien Nguyen
William Guicquero
Gilles Sicard
MQ
69
1
0
17 Jan 2025
Histogram-Equalized Quantization for logic-gated Residual Neural Networks
Van Thien Nguyen
William Guicquero
Gilles Sicard
MQ
35
1
0
10 Jan 2025
Self-Masking Networks for Unsupervised Adaptation
Alfonso Taboada Warmerdam
Mathilde Caron
Yuki M. Asano
39
1
0
11 Sep 2024
Quality Scalable Quantization Methodology for Deep Learning on Edge
S. Khaliq
Rehan Hafiz
MQ
28
1
0
15 Jul 2024
TerDiT: Ternary Diffusion Models with Transformers
Xudong Lu
Aojun Zhou
Ziyi Lin
Qi Liu
Yuhui Xu
Renrui Zhang
Yafei Wen
Shuai Ren
Peng Gao
Junchi Yan
MQ
37
2
0
23 May 2024
AdaQAT: Adaptive Bit-Width Quantization-Aware Training
Cédric Gernigon
Silviu-Ioan Filip
Olivier Sentieys
Clément Coggiola
Mickael Bruno
23
2
0
22 Apr 2024
Better Schedules for Low Precision Training of Deep Neural Networks
Cameron R. Wolfe
Anastasios Kyrillidis
40
1
0
04 Mar 2024
PLUM: Improving Inference Efficiency By Leveraging Repetition-Sparsity Trade-Off
Sachit Kuhar
Yash Jain
Alexey Tumanov
MQ
52
0
0
04 Dec 2023
AutoFHE: Automated Adaption of CNNs for Efficient Evaluation over FHE
Wei Ao
Vishnu Naresh Boddeti
AAML
17
18
0
12 Oct 2023
Learning Discrete Weights and Activations Using the Local Reparameterization Trick
G. Berger
Aviv Navon
Ethan Fetaya
MQ
20
0
0
04 Jul 2023
Data-Free Quantization via Mixed-Precision Compensation without Fine-Tuning
Jun Chen
Shipeng Bai
Tianxin Huang
Mengmeng Wang
Guanzhong Tian
Y. Liu
MQ
34
18
0
02 Jul 2023
Evaluation Metrics for DNNs Compression
Abanoub Ghobrial
S. Budgett
Dieter Balemans
Hamid Asgari
Philippe Reiter
Kerstin Eder
22
1
0
18 May 2023
AutoQNN: An End-to-End Framework for Automatically Quantizing Neural Networks
Cheng Gong
Ye Lu
Surong Dai
Deng Qian
Chenkun Du
Tao Li
MQ
24
0
0
07 Apr 2023
Learning Discretized Neural Networks under Ricci Flow
Jun Chen
Han Chen
Mengmeng Wang
Guang Dai
Ivor W. Tsang
Y. Liu
13
2
0
07 Feb 2023
Hyperspherical Quantization: Toward Smaller and More Accurate Models
Dan Liu
X. Chen
Chen-li Ma
Xue Liu
MQ
20
3
0
24 Dec 2022
CSQ: Growing Mixed-Precision Quantization Scheme with Bi-level Continuous Sparsification
Lirui Xiao
Huanrui Yang
Zhen Dong
Kurt Keutzer
Li Du
Shanghang Zhang
MQ
24
10
0
06 Dec 2022
Signed Binary Weight Networks
Sachit Kuhar
Alexey Tumanov
Judy Hoffman
MQ
11
1
0
25 Nov 2022
AskewSGD : An Annealed interval-constrained Optimisation method to train Quantized Neural Networks
Louis Leconte
S. Schechtman
Eric Moulines
27
4
0
07 Nov 2022
Seeking Interpretability and Explainability in Binary Activated Neural Networks
Benjamin Leblanc
Pascal Germain
FAtt
27
1
0
07 Sep 2022
AdaBin: Improving Binary Neural Networks with Adaptive Binary Sets
Zhaopeng Tu
Xinghao Chen
Pengju Ren
Yunhe Wang
MQ
32
54
0
17 Aug 2022
Mixed-Precision Neural Networks: A Survey
M. Rakka
M. Fouda
Pramod P. Khargonekar
Fadi J. Kurdahi
MQ
18
11
0
11 Aug 2022
Combinatorial optimization for low bit-width neural networks
Hanxu Zhou
Aida Ashrafi
Matthew B. Blaschko
MQ
19
0
0
04 Jun 2022
ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers
Z. Yao
Reza Yazdani Aminabadi
Minjia Zhang
Xiaoxia Wu
Conglong Li
Yuxiong He
VLM
MQ
39
438
0
04 Jun 2022
LilNetX: Lightweight Networks with EXtreme Model Compression and Structured Sparsification
Sharath Girish
Kamal Gupta
Saurabh Singh
Abhinav Shrivastava
26
11
0
06 Apr 2022
EAutoDet: Efficient Architecture Search for Object Detection
Xiaoxing Wang
Jiale Lin
Junchi Yan
Juanping Zhao
Xiaokang Yang
22
23
0
21 Mar 2022
FlexBlock: A Flexible DNN Training Accelerator with Multi-Mode Block Floating Point Support
Seock-Hwan Noh
Jahyun Koo
Seunghyun Lee
Jongse Park
Jaeha Kung
AI4CE
11
17
0
13 Mar 2022
Standard Deviation-Based Quantization for Deep Neural Networks
Amir Ardakani
A. Ardakani
B. Meyer
J. Clark
W. Gross
MQ
35
1
0
24 Feb 2022
Bitwidth Heterogeneous Federated Learning with Progressive Weight Dequantization
Jaehong Yoon
Geondo Park
Wonyong Jeong
Sung Ju Hwang
FedML
15
19
0
23 Feb 2022
Distilled Neural Networks for Efficient Learning to Rank
F. M. Nardini
Cosimo Rulli
Salvatore Trani
Rossano Venturini
FedML
16
16
0
22 Feb 2022
Bit-wise Training of Neural Network Weights
Cristian Ivan
MQ
16
0
0
19 Feb 2022
The Ecological Footprint of Neural Machine Translation Systems
D. Shterionov
Eva Vanmassenhove
24
3
0
04 Feb 2022
Signing the Supermask: Keep, Hide, Invert
Nils Koster
O. Grothe
Achim Rettinger
23
10
0
31 Jan 2022
Elastic-Link for Binarized Neural Network
Jie Hu
Ziheng Wu
Vince Tan
Zhilin Lu
Mengze Zeng
Enhua Wu
MQ
16
6
0
19 Dec 2021
Neural Network Quantization for Efficient Inference: A Survey
Olivia Weng
MQ
14
22
0
08 Dec 2021
Mixed Precision of Quantization of Transformer Language Models for Speech Recognition
Junhao Xu
Shoukang Hu
Jianwei Yu
Xunying Liu
Helen M. Meng
MQ
30
15
0
29 Nov 2021
Toward Compact Parameter Representations for Architecture-Agnostic Neural Network Compression
Yuezhou Sun
Wenlong Zhao
Lijun Zhang
Xiao Liu
Hui Guan
Matei A. Zaharia
21
0
0
19 Nov 2021
BNAS v2: Learning Architectures for Binary Networks with Empirical Improvements
Dahyun Kim
Kunal Pratap Singh
Jonghyun Choi
MQ
38
7
0
16 Oct 2021
Towards Mixed-Precision Quantization of Neural Networks via Constrained Optimization
Weihan Chen
Peisong Wang
Jian Cheng
MQ
31
61
0
13 Oct 2021
CBP: Backpropagation with constraint on weight precision using a pseudo-Lagrange multiplier method
Guhyun Kim
D. Jeong
MQ
34
2
0
06 Oct 2021
Communication-Efficient Federated Learning with Binary Neural Networks
YuZhi Yang
Zhaoyang Zhang
Qianqian Yang
FedML
16
31
0
05 Oct 2021
Towards Efficient Post-training Quantization of Pre-trained Language Models
Haoli Bai
Lu Hou
Lifeng Shang
Xin Jiang
Irwin King
M. Lyu
MQ
71
47
0
30 Sep 2021
Prune Your Model Before Distill It
Jinhyuk Park
Albert No
VLM
38
27
0
30 Sep 2021
Content-Aware Convolutional Neural Networks
Yong Guo
Yaofo Chen
Mingkui Tan
K. Jia
Jian Chen
Jingdong Wang
27
8
0
30 Jun 2021
Reward-Based 1-bit Compressed Federated Distillation on Blockchain
Leon Witt
Usama Zafar
KuoYeh Shen
Felix Sattler
Dan Li
Wojciech Samek
FedML
24
4
0
27 Jun 2021
Efficient Deep Learning: A Survey on Making Deep Learning Models Smaller, Faster, and Better
Gaurav Menghani
VLM
MedIm
23
360
0
16 Jun 2021
1
2
Next