ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2002.03090
  4. Cited By
BitPruning: Learning Bitlengths for Aggressive and Accurate Quantization
v1v2 (latest)

BitPruning: Learning Bitlengths for Aggressive and Accurate Quantization

International Symposium on Circuits and Systems (ISCAS), 2020
8 February 2020
Milovs Nikolić
G. B. Hacene
Ciaran Bannon
Alberto Delmas Lascorz
Matthieu Courbariaux
Yoshua Bengio
Vincent Gripon
Andreas Moshovos
    MQ
ArXiv (abs)PDFHTML

Papers citing "BitPruning: Learning Bitlengths for Aggressive and Accurate Quantization"

16 / 16 papers shown
AdaQAT: Adaptive Bit-Width Quantization-Aware Training
AdaQAT: Adaptive Bit-Width Quantization-Aware Training
Cédric Gernigon
Silviu-Ioan Filip
Olivier Sentieys
Clément Coggiola
Mickael Bruno
223
6
0
22 Apr 2024
CNN-Based Equalization for Communications: Achieving Gigabit Throughput
  with a Flexible FPGA Hardware Architecture
CNN-Based Equalization for Communications: Achieving Gigabit Throughput with a Flexible FPGA Hardware Architecture
Jonas Ney
C. Füllner
V. Lauinger
Laurent Schmalen
Sebastian Randel
Norbert Wehn
212
1
0
22 Apr 2024
Free Bits: Latency Optimization of Mixed-Precision Quantized Neural
  Networks on the Edge
Free Bits: Latency Optimization of Mixed-Precision Quantized Neural Networks on the EdgeInternational Conference on Artificial Intelligence Circuits and Systems (ICAICS), 2023
Georg Rutishauser
Francesco Conti
Luca Benini
MQ
280
5
0
06 Jul 2023
Unsupervised ANN-Based Equalizer and Its Trainable FPGA Implementation
Unsupervised ANN-Based Equalizer and Its Trainable FPGA Implementation
Jonas Ney
V. Lauinger
Laurent Schmalen
Norbert Wehn
253
6
0
14 Apr 2023
Efficient and Effective Methods for Mixed Precision Neural Network
  Quantization for Faster, Energy-efficient Inference
Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference
Deepika Bablani
J. McKinstry
S. K. Esser
R. Appuswamy
D. Modha
MQ
325
10
0
30 Jan 2023
FullPack: Full Vector Utilization for Sub-Byte Quantized Inference on
  General Purpose CPUs
FullPack: Full Vector Utilization for Sub-Byte Quantized Inference on General Purpose CPUs
Hossein Katebi
Navidreza Asadi
M. Goudarzi
MQ
174
1
0
13 Nov 2022
SDQ: Stochastic Differentiable Quantization with Mixed Precision
SDQ: Stochastic Differentiable Quantization with Mixed PrecisionInternational Conference on Machine Learning (ICML), 2022
Xijie Huang
Zhiqiang Shen
Shichao Li
Zechun Liu
Xianghong Hu
Jeffry Wicaksana
Eric P. Xing
Kwang-Ting Cheng
MQ
473
48
0
09 Jun 2022
A Low Memory Footprint Quantized Neural Network for Depth Completion of
  Very Sparse Time-of-Flight Depth Maps
A Low Memory Footprint Quantized Neural Network for Depth Completion of Very Sparse Time-of-Flight Depth Maps
Xiao-Yan Jiang
V. Cambareri
Gianluca Agresti
C. Ugwu
Adriano Simonetto
Fabien Cardinaux
Pietro Zanuttigh
3DVMQ
207
11
0
25 May 2022
A Silicon Photonic Accelerator for Convolutional Neural Networks with
  Heterogeneous Quantization
A Silicon Photonic Accelerator for Convolutional Neural Networks with Heterogeneous QuantizationACM Great Lakes Symposium on VLSI (GLSVLSI), 2022
Febin P. Sunny
Mahdi Nikdast
S. Pasricha
MQ
165
25
0
17 May 2022
A Comprehensive Survey on Model Quantization for Deep Neural Networks in
  Image Classification
A Comprehensive Survey on Model Quantization for Deep Neural Networks in Image ClassificationACM Transactions on Intelligent Systems and Technology (ACM TIST), 2022
Babak Rokh
A. Azarpeyvand
Alireza Khanteymoori
MQ
512
205
0
14 May 2022
Schrödinger's FP: Dynamic Adaptation of Floating-Point Containers for
  Deep Learning Training
Schrödinger's FP: Dynamic Adaptation of Floating-Point Containers for Deep Learning Training
Milovs Nikolić
Enrique Torres Sanchez
Jia-Hui Wang
Ali Hadi Zadeh
Mostafa Mahmoud
Ameer Abdelhadi
Kareem Ibrahim
Andreas Moshovos
MQ
318
1
0
28 Apr 2022
APack: Off-Chip, Lossless Data Compression for Efficient Deep Learning
  Inference
APack: Off-Chip, Lossless Data Compression for Efficient Deep Learning Inference
Alberto Delmas Lascorz
Mostafa Mahmoud
Andreas Moshovos
MQ
175
1
0
21 Jan 2022
Quantization and Deployment of Deep Neural Networks on Microcontrollers
Quantization and Deployment of Deep Neural Networks on MicrocontrollersItalian National Conference on Sensors (INS), 2021
Pierre-Emmanuel Novac
G. B. Hacene
Alain Pegatoquet
Benoit Miramond
Vincent Gripon
MQ
244
166
0
27 May 2021
DNN Quantization with Attention
DNN Quantization with Attention
G. B. Hacene
Lukas Mauch
Stefan Uhlich
Fabien Cardinaux
MQ
198
2
0
24 Mar 2021
DecisiveNets: Training Deep Associative Memories to Solve Complex
  Machine Learning Problems
DecisiveNets: Training Deep Associative Memories to Solve Complex Machine Learning Problems
Vincent Gripon
Carlos Lassance
G. B. Hacene
163
0
0
02 Dec 2020
FracBits: Mixed Precision Quantization via Fractional Bit-Widths
FracBits: Mixed Precision Quantization via Fractional Bit-Widths
Linjie Yang
Qing Jin
MQ
299
93
0
04 Jul 2020
1
Page 1 of 1