ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2203.05740
  4. Cited By
QDrop: Randomly Dropping Quantization for Extremely Low-bit
  Post-Training Quantization
v1v2 (latest)

QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization

International Conference on Learning Representations (ICLR), 2022
11 March 2022
Xiuying Wei
Yazhe Niu
Yuhang Li
Xianglong Liu
F. Yu
    MQVLM
ArXiv (abs)PDFHTMLGithub (122★)

Papers citing "QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization"

50 / 124 papers shown
Advancing Multimodal Large Language Models with Quantization-Aware Scale
  Learning for Efficient Adaptation
Advancing Multimodal Large Language Models with Quantization-Aware Scale Learning for Efficient AdaptationACM Multimedia (MM), 2024
Jingjing Xie
Yuxin Zhang
Mingbao Lin
Liujuan Cao
Rongrong Ji
MQ
177
14
0
07 Aug 2024
DopQ-ViT: Towards Distribution-Friendly and Outlier-Aware Post-Training Quantization for Vision Transformers
DopQ-ViT: Towards Distribution-Friendly and Outlier-Aware Post-Training Quantization for Vision Transformers
Lianwei Yang
Haisong Gong
Haokun Lin
Yichen Wu
Zhenan Sun
Qingyi Gu
MQ
479
11
0
06 Aug 2024
Temporal Feature Matters: A Framework for Diffusion Model Quantization
Temporal Feature Matters: A Framework for Diffusion Model QuantizationIEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2024
Yushi Huang
Yazhe Niu
Xianglong Liu
Jing Liu
Yuhang Li
Jiwen Lu
Dacheng Tao
MQDiffM
652
4
0
28 Jul 2024
Compensate Quantization Errors+: Quantized Models Are Inquisitive Learners
Compensate Quantization Errors+: Quantized Models Are Inquisitive Learners
Yifei Gao
Jie Ou
Lei Wang
Fanhua Shang
Jaji Wu
MQ
369
0
0
22 Jul 2024
MetaAug: Meta-Data Augmentation for Post-Training Quantization
MetaAug: Meta-Data Augmentation for Post-Training Quantization
Cuong Pham
Hoang Anh Dung
Cuong C. Nguyen
Trung Le
Dinh Q. Phung
Gustavo Carneiro
Thanh-Toan Do
MQ
237
1
0
20 Jul 2024
AdaLog: Post-Training Quantization for Vision Transformers with Adaptive
  Logarithm Quantizer
AdaLog: Post-Training Quantization for Vision Transformers with Adaptive Logarithm Quantizer
Zhuguanyu Wu
Jiaxin Chen
Hanwen Zhong
Di Huang
Yun Wang
MQ
340
23
0
17 Jul 2024
QVD: Post-training Quantization for Video Diffusion Models
QVD: Post-training Quantization for Video Diffusion Models
Shilong Tian
Hong Chen
Chengtao Lv
Yu Liu
Jinyang Guo
Xianglong Liu
Shengxi Li
Hao Yang
Tao Xie
VGenMQ
280
12
0
16 Jul 2024
NITRO-D: Native Integer-only Training of Deep Convolutional Neural Networks
NITRO-D: Native Integer-only Training of Deep Convolutional Neural Networks
Alberto Pirillo
Luca Colombo
Manuel Roveri
MQ
413
1
0
16 Jul 2024
LRQ: Optimizing Post-Training Quantization for Large Language Models by Learning Low-Rank Weight-Scaling Matrices
LRQ: Optimizing Post-Training Quantization for Large Language Models by Learning Low-Rank Weight-Scaling Matrices
Jung Hyun Lee
Jeonghoon Kim
J. Yang
S. Kwon
Eunho Yang
Kang Min Yoo
Dongsoo Lee
MQ
345
5
0
16 Jul 2024
ISQuant: apply squant to the real deployment
ISQuant: apply squant to the real deployment
Dezan Zhao
MQ
193
0
0
05 Jul 2024
ADFQ-ViT: Activation-Distribution-Friendly Post-Training Quantization
  for Vision Transformers
ADFQ-ViT: Activation-Distribution-Friendly Post-Training Quantization for Vision Transformers
Yanfeng Jiang
Ning Sun
Xueshuo Xie
Fei Yang
Tao Li
MQ
303
11
0
03 Jul 2024
Compensate Quantization Errors: Make Weights Hierarchical to Compensate
  Each Other
Compensate Quantization Errors: Make Weights Hierarchical to Compensate Each Other
Yifei Gao
Jie Ou
Lei Wang
Yuting Xiao
Zhiyuan Xiang
Ruiting Dai
Jun Cheng
MQ
196
5
0
24 Jun 2024
MGRQ: Post-Training Quantization For Vision Transformer With Mixed
  Granularity Reconstruction
MGRQ: Post-Training Quantization For Vision Transformer With Mixed Granularity Reconstruction
Lianwei Yang
Zhikai Li
Junrui Xiao
Haisong Gong
Qingyi Gu
MQ
193
6
0
13 Jun 2024
Asymptotic Unbiased Sample Sampling to Speed Up Sharpness-Aware Minimization
Asymptotic Unbiased Sample Sampling to Speed Up Sharpness-Aware Minimization
Jiaxin Deng
Junbiao Pang
Baochang Zhang
473
4
0
12 Jun 2024
MagR: Weight Magnitude Reduction for Enhancing Post-Training
  Quantization
MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Aozhong Zhang
Naigang Wang
Yanxia Deng
Xin Li
Zi Yang
Penghang Yin
MQ
227
16
0
02 Jun 2024
LCQ: Low-Rank Codebook based Quantization for Large Language Models
LCQ: Low-Rank Codebook based Quantization for Large Language Models
Wen-Pu Cai
Wu-Jun Li
Wu-Jun Li
MQ
338
0
0
31 May 2024
Information Entropy Guided Height-aware Histogram for Quantization-friendly Pillar Feature Encoder
Information Entropy Guided Height-aware Histogram for Quantization-friendly Pillar Feature Encoder
Sifan Zhou
Zhihang Yuan
Dawei Yang
Ziyu Zhao
Yan Chen
Xing Hu
Ziyu Zhao
Xiaobo Lu
617
2
0
29 May 2024
I-LLM: Efficient Integer-Only Inference for Fully-Quantized Low-Bit
  Large Language Models
I-LLM: Efficient Integer-Only Inference for Fully-Quantized Low-Bit Large Language Models
Yan Chen
Yuan Cheng
Dawei Yang
Zhihang Yuan
Jiangyong Yu
Chen Xu
Sifan Zhou
MQ
409
16
0
28 May 2024
PTQ4DiT: Post-training Quantization for Diffusion Transformers
PTQ4DiT: Post-training Quantization for Diffusion Transformers
Junyi Wu
Haoxuan Wang
Yuzhang Shang
Mubarak Shah
Yan Yan
MQ
306
42
0
25 May 2024
Nearest is Not Dearest: Towards Practical Defense against
  Quantization-conditioned Backdoor Attacks
Nearest is Not Dearest: Towards Practical Defense against Quantization-conditioned Backdoor Attacks
Boheng Li
Yishuo Cai
Haowei Li
Feng Xue
Zhifeng Li
Yiming Li
MQAAML
260
28
0
21 May 2024
Selective Focus: Investigating Semantics Sensitivity in Post-training
  Quantization for Lane Detection
Selective Focus: Investigating Semantics Sensitivity in Post-training Quantization for Lane DetectionAAAI Conference on Artificial Intelligence (AAAI), 2024
Yunqian Fan
Xiuying Wei
Yazhe Niu
Yuqing Ma
Xiangguo Zhang
Qi Zhang
Xianglong Liu
MQ
218
3
0
10 May 2024
Fast and Controllable Post-training Sparsity: Learning Optimal Sparsity
  Allocation with Global Constraint in Minutes
Fast and Controllable Post-training Sparsity: Learning Optimal Sparsity Allocation with Global Constraint in MinutesAAAI Conference on Artificial Intelligence (AAAI), 2024
Yazhe Niu
Yang Yong
Zining Wang
Jinyang Guo
Xiuying Wei
Yuqing Ma
Xianglong Liu
194
7
0
09 May 2024
LLMC: Benchmarking Large Language Model Quantization with a Versatile
  Compression Toolkit
LLMC: Benchmarking Large Language Model Quantization with a Versatile Compression ToolkitConference on Empirical Methods in Natural Language Processing (EMNLP), 2024
Yazhe Niu
Yang Yong
Shiqiao Gu
Yushi Huang
Chentao Lv
Yunchen Zhang
Xianglong Liu
Dacheng Tao
MQ
342
22
0
09 May 2024
PTQ4SAM: Post-Training Quantization for Segment Anything
PTQ4SAM: Post-Training Quantization for Segment AnythingComputer Vision and Pattern Recognition (CVPR), 2024
Chengtao Lv
Hong Chen
Jinyang Guo
Yifu Ding
Xianglong Liu
VLMMQ
198
29
0
06 May 2024
Torch2Chip: An End-to-end Customizable Deep Neural Network Compression
  and Deployment Toolkit for Prototype Hardware Accelerator Design
Torch2Chip: An End-to-end Customizable Deep Neural Network Compression and Deployment Toolkit for Prototype Hardware Accelerator DesignConference on Machine Learning and Systems (MLSys), 2024
Jian Meng
Yuan Liao
Anupreetham Anupreetham
Ahmed Hassan
Shixing Yu
Han-Sok Suh
Xiaofeng Hu
Jae-sun Seo
MQ
211
2
0
02 May 2024
Frame Quantization of Neural Networks
Frame Quantization of Neural Networks
Wojciech Czaja
Sanghoon Na
210
1
0
11 Apr 2024
Instance-Aware Group Quantization for Vision Transformers
Instance-Aware Group Quantization for Vision Transformers
Jaehyeon Moon
Jeimin Jeon
Junyong Cheon
Bumsub Ham
MQViT
251
15
0
01 Apr 2024
AffineQuant: Affine Transformation Quantization for Large Language
  Models
AffineQuant: Affine Transformation Quantization for Large Language Models
Yuexiao Ma
Huixia Li
Xiawu Zheng
Feng Ling
Xuefeng Xiao
Rui Wang
Shilei Wen
Jiayi Ji
Rongrong Ji
MQ
251
42
0
19 Mar 2024
COMQ: A Backpropagation-Free Algorithm for Post-Training Quantization
COMQ: A Backpropagation-Free Algorithm for Post-Training QuantizationIEEE Access (IEEE Access), 2024
Aozhong Zhang
Zi Yang
Naigang Wang
Yingyong Qin
Jack Xin
Xin Li
Penghang Yin
VLMMQ
206
12
0
11 Mar 2024
Towards Accurate Post-training Quantization for Reparameterized Models
Towards Accurate Post-training Quantization for Reparameterized Models
Luoming Zhang
Yefei He
Wen Fei
Zhenyu Lou
Weijia Wu
YangWei Ying
Hong Zhou
MQ
193
1
0
25 Feb 2024
Effective Gradient Sample Size via Variation Estimation for Accelerating
  Sharpness aware Minimization
Effective Gradient Sample Size via Variation Estimation for Accelerating Sharpness aware Minimization
Jiaxin Deng
Junbiao Pang
Baochang Zhang
Tian Wang
209
1
0
24 Feb 2024
QuEST: Low-bit Diffusion Model Quantization via Efficient Selective Finetuning
QuEST: Low-bit Diffusion Model Quantization via Efficient Selective Finetuning
Haoxuan Wang
Yuzhang Shang
Zhihang Yuan
Junyi Wu
Junchi Yan
Yan Yan
MQDiffM
394
43
0
06 Feb 2024
LiDAR-PTQ: Post-Training Quantization for Point Cloud 3D Object
  Detection
LiDAR-PTQ: Post-Training Quantization for Point Cloud 3D Object DetectionInternational Conference on Learning Representations (ICLR), 2024
Sifan Zhou
Liang Li
Xinyu Zhang
Bo Zhang
Shipeng Bai
Miao Sun
Ziyu Zhao
Xiaobo Lu
Xiangxiang Chu
MQ
206
31
0
29 Jan 2024
GenQ: Quantization in Low Data Regimes with Generative Synthetic Data
GenQ: Quantization in Low Data Regimes with Generative Synthetic DataEuropean Conference on Computer Vision (ECCV), 2023
Yuhang Li
Youngeun Kim
Donghyun Lee
Souvik Kundu
Priyadarshini Panda
MQ
334
6
0
07 Dec 2023
TFMQ-DM: Temporal Feature Maintenance Quantization for Diffusion Models
TFMQ-DM: Temporal Feature Maintenance Quantization for Diffusion ModelsComputer Vision and Pattern Recognition (CVPR), 2023
Yushi Huang
Yazhe Niu
Jing Liu
Tianlong Chen
Xianglong Liu
DiffMMQ
279
65
0
27 Nov 2023
I&S-ViT: An Inclusive & Stable Method for Pushing the Limit of Post-Training ViTs Quantization
I&S-ViT: An Inclusive & Stable Method for Pushing the Limit of Post-Training ViTs Quantization
Mingliang Xu
Jiawei Hu
Mingbao Lin
Mengzhao Chen
Rongrong Ji
MQ
408
14
0
16 Nov 2023
Quantized Distillation: Optimizing Driver Activity Recognition Models
  for Resource-Constrained Environments
Quantized Distillation: Optimizing Driver Activity Recognition Models for Resource-Constrained EnvironmentsIEEE/RJS International Conference on Intelligent RObots and Systems (IROS), 2023
Calvin Tanama
Kunyu Peng
Zdravko Marinov
Rainer Stiefelhagen
Alina Roitberg
198
2
0
10 Nov 2023
LLM-FP4: 4-Bit Floating-Point Quantized Transformers
LLM-FP4: 4-Bit Floating-Point Quantized TransformersConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Shih-yang Liu
Zechun Liu
Xijie Huang
Pingcheng Dong
Kwang-Ting Cheng
MQ
239
92
0
25 Oct 2023
TEQ: Trainable Equivalent Transformation for Quantization of LLMs
TEQ: Trainable Equivalent Transformation for Quantization of LLMs
Wenhua Cheng
Yiyang Cai
Kaokao Lv
Haihao Shen
MQ
281
10
0
17 Oct 2023
QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large
  Language Models
QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language ModelsInternational Conference on Learning Representations (ICLR), 2023
Jing Liu
Yazhe Niu
Xiuying Wei
Zhiwei Dong
Jianfei Cai
Bohan Zhuang
MQ
299
67
0
12 Oct 2023
Dual Grained Quantization: Efficient Fine-Grained Quantization for LLM
Dual Grained Quantization: Efficient Fine-Grained Quantization for LLM
Luoming Zhang
Wen Fei
Weijia Wu
Yefei He
Zhenyu Lou
Hong Zhou
MQ
212
5
0
07 Oct 2023
EfficientDM: Efficient Quantization-Aware Fine-Tuning of Low-Bit
  Diffusion Models
EfficientDM: Efficient Quantization-Aware Fine-Tuning of Low-Bit Diffusion ModelsInternational Conference on Learning Representations (ICLR), 2023
Yefei He
Jing Liu
Weijia Wu
Hong Zhou
Bohan Zhuang
DiffMMQ
522
68
0
05 Oct 2023
Network Memory Footprint Compression Through Jointly Learnable Codebooks
  and Mappings
Network Memory Footprint Compression Through Jointly Learnable Codebooks and MappingsInternational Conference on Learning Representations (ICLR), 2023
Vittorio Giammarino
Arnaud Dapogny
Kévin Bailly
MQ
187
2
0
29 Sep 2023
EPTQ: Enhanced Post-Training Quantization via Label-Free Hessian
EPTQ: Enhanced Post-Training Quantization via Label-Free Hessian
Ofir Gordon
H. Habi
Arnon Netzer
MQ
235
1
0
20 Sep 2023
Optimize Weight Rounding via Signed Gradient Descent for the
  Quantization of LLMs
Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Wenhua Cheng
Weiwei Zhang
Haihao Shen
Yiyang Cai
Xin He
Kaokao Lv
Yi. Liu
MQ
507
33
0
11 Sep 2023
Gradient-Based Post-Training Quantization: Challenging the Status Quo
Gradient-Based Post-Training Quantization: Challenging the Status Quo
Edouard Yvinec
Arnaud Dapogny
Kévin Bailly
MQ
223
1
0
15 Aug 2023
EQ-Net: Elastic Quantization Neural Networks
EQ-Net: Elastic Quantization Neural NetworksIEEE International Conference on Computer Vision (ICCV), 2023
Ke Xu
Lei Han
Ye Tian
Shangshang Yang
Xingyi Zhang
MQ
348
18
0
15 Aug 2023
NUPES : Non-Uniform Post-Training Quantization via Power Exponent Search
NUPES : Non-Uniform Post-Training Quantization via Power Exponent SearchIEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2023
Edouard Yvinec
Arnaud Dapogny
Kévin Bailly
MQ
174
9
0
10 Aug 2023
Lossy and Lossless (L$^2$) Post-training Model Size Compression
Lossy and Lossless (L2^22) Post-training Model Size CompressionIEEE International Conference on Computer Vision (ICCV), 2023
Yumeng Shi
Shihao Bai
Xiuying Wei
Yazhe Niu
Jianlei Yang
189
5
0
08 Aug 2023
Squeezing Large-Scale Diffusion Models for Mobile
Squeezing Large-Scale Diffusion Models for Mobile
Jiwoong Choi
Minkyu Kim
Daehyun Ahn
Taesu Kim
Yulhwa Kim
Do-Hyun Jo
H. Jeon
Jae-Joon Kim
Hyungjun Kim
206
13
0
03 Jul 2023
Previous
123
Next