ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2112.13843
  4. Cited By
BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of
  DNNs from Scratch

BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of DNNs from Scratch

24 December 2021
Souvik Kundu
Shikai Wang
Qirui Sun
P. Beerel
Massoud Pedram
    MQ
ArXivPDFHTML

Papers citing "BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of DNNs from Scratch"

13 / 13 papers shown
Title
OuroMamba: A Data-Free Quantization Framework for Vision Mamba Models
Akshat Ramachandran
Mingyu Lee
Huan Xu
Souvik Kundu
Tushar Krishna
MQ
51
1
0
13 Mar 2025
FAMES: Fast Approximate Multiplier Substitution for Mixed-Precision
  Quantized DNNs--Down to 2 Bits!
FAMES: Fast Approximate Multiplier Substitution for Mixed-Precision Quantized DNNs--Down to 2 Bits!
Yi Ren
Ruge Xu
Xinfei Guo
Weikang Qian
MQ
69
0
0
27 Nov 2024
CLAMP-ViT: Contrastive Data-Free Learning for Adaptive Post-Training
  Quantization of ViTs
CLAMP-ViT: Contrastive Data-Free Learning for Adaptive Post-Training Quantization of ViTs
Akshat Ramachandran
Souvik Kundu
Tushar Krishna
MQ
32
9
0
07 Jul 2024
Block Selective Reprogramming for On-device Training of Vision
  Transformers
Block Selective Reprogramming for On-device Training of Vision Transformers
Sreetama Sarkar
Souvik Kundu
Kai Zheng
P. Beerel
37
2
0
25 Mar 2024
Value-Driven Mixed-Precision Quantization for Patch-Based Inference on
  Microcontrollers
Value-Driven Mixed-Precision Quantization for Patch-Based Inference on Microcontrollers
Wei Tao
Shenglin He
Kai Lu
Xiaoyang Qu
Guokuan Li
Jiguang Wan
Jianzong Wang
Jing Xiao
MQ
23
0
0
24 Jan 2024
LQ-LoRA: Low-rank Plus Quantized Matrix Decomposition for Efficient
  Language Model Finetuning
LQ-LoRA: Low-rank Plus Quantized Matrix Decomposition for Efficient Language Model Finetuning
Han Guo
P. Greengard
Eric P. Xing
Yoon Kim
MQ
36
43
0
20 Nov 2023
Learning to Linearize Deep Neural Networks for Secure and Efficient
  Private Inference
Learning to Linearize Deep Neural Networks for Secure and Efficient Private Inference
Souvik Kundu
Shun Lu
Yuke Zhang
Jacqueline Liu
P. Beerel
8
29
0
23 Jan 2023
Sparse Mixture Once-for-all Adversarial Training for Efficient In-Situ
  Trade-Off Between Accuracy and Robustness of DNNs
Sparse Mixture Once-for-all Adversarial Training for Efficient In-Situ Trade-Off Between Accuracy and Robustness of DNNs
Souvik Kundu
Sairam Sundaresan
S. N. Sridhar
Shunlin Lu
Han Tang
P. Beerel
AAML
MoE
44
4
0
27 Dec 2022
FIT: A Metric for Model Sensitivity
FIT: A Metric for Model Sensitivity
Ben Zandonati
Adrian Alan Pol
M. Pierini
Olya Sirkin
Tal Kopetz
MQ
24
8
0
16 Oct 2022
Self-Attentive Pooling for Efficient Deep Learning
Self-Attentive Pooling for Efficient Deep Learning
Fang Chen
Gourav Datta
Souvik Kundu
P. Beerel
79
6
0
16 Sep 2022
RBNN: Memory-Efficient Reconfigurable Deep Binary Neural Network with IP
  Protection for Internet of Things
RBNN: Memory-Efficient Reconfigurable Deep Binary Neural Network with IP Protection for Internet of Things
Huming Qiu
Hua Ma
Zhi-Li Zhang
Yifeng Zheng
Anmin Fu
Pan Zhou
Yansong Gao
Derek Abbott
S. Al-Sarawi
MQ
19
9
0
09 May 2021
Activation Density based Mixed-Precision Quantization for Energy
  Efficient Neural Networks
Activation Density based Mixed-Precision Quantization for Energy Efficient Neural Networks
Karina Vasquez
Yeshwanth Venkatesha
Abhiroop Bhattacharjee
Abhishek Moitra
Priyadarshini Panda
MQ
35
15
0
12 Jan 2021
Incremental Network Quantization: Towards Lossless CNNs with
  Low-Precision Weights
Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights
Aojun Zhou
Anbang Yao
Yiwen Guo
Lin Xu
Yurong Chen
MQ
319
1,049
0
10 Feb 2017
1