ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1912.08883
  4. Cited By
Adaptive Loss-aware Quantization for Multi-bit Networks

Adaptive Loss-aware Quantization for Multi-bit Networks

18 December 2019
Zhongnan Qu
Zimu Zhou
Yun Cheng
Lothar Thiele
    MQ
ArXivPDFHTML

Papers citing "Adaptive Loss-aware Quantization for Multi-bit Networks"

26 / 26 papers shown
Title
Learning from Loss Landscape: Generalizable Mixed-Precision Quantization via Adaptive Sharpness-Aware Gradient Aligning
Learning from Loss Landscape: Generalizable Mixed-Precision Quantization via Adaptive Sharpness-Aware Gradient Aligning
Lianbo Ma
Jianlun Ma
Yuee Zhou
Guoyang Xie
Qiang He
Zhichao Lu
MQ
36
0
0
08 May 2025
Radio: Rate-Distortion Optimization for Large Language Model Compression
Radio: Rate-Distortion Optimization for Large Language Model Compression
Sean I. Young
MQ
21
0
0
05 May 2025
Channel-Wise Mixed-Precision Quantization for Large Language Models
Channel-Wise Mixed-Precision Quantization for Large Language Models
Zihan Chen
Bike Xie
Jundong Li
Cong Shen
MQ
27
2
0
16 Oct 2024
Error Diffusion: Post Training Quantization with Block-Scaled Number
  Formats for Neural Networks
Error Diffusion: Post Training Quantization with Block-Scaled Number Formats for Neural Networks
Alireza Khodamoradi
K. Denolf
Eric Dellinger
MQ
26
0
0
15 Oct 2024
Foundations of Large Language Model Compression -- Part 1: Weight
  Quantization
Foundations of Large Language Model Compression -- Part 1: Weight Quantization
Sean I. Young
MQ
35
1
0
03 Sep 2024
From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of
  Deep Neural Networks
From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of Deep Neural Networks
Xue Geng
Zhe Wang
Chunyun Chen
Qing Xu
Kaixin Xu
...
Zhenghua Chen
M. Aly
Jie Lin
Min-man Wu
Xiaoli Li
33
1
0
09 May 2024
One-Step Forward and Backtrack: Overcoming Zig-Zagging in Loss-Aware
  Quantization Training
One-Step Forward and Backtrack: Overcoming Zig-Zagging in Loss-Aware Quantization Training
Lianbo Ma
Yuee Zhou
Jianlun Ma
Guo-Ding Yu
Qing Li
MQ
15
1
0
30 Jan 2024
AccEPT: An Acceleration Scheme for Speeding Up Edge Pipeline-parallel
  Training
AccEPT: An Acceleration Scheme for Speeding Up Edge Pipeline-parallel Training
Yuhao Chen
Yuxuan Yan
Qianqian Yang
Yuanchao Shu
Shibo He
Zhiguo Shi
Jiming Chen
27
0
0
10 Nov 2023
Early-Exit with Class Exclusion for Efficient Inference of Neural
  Networks
Early-Exit with Class Exclusion for Efficient Inference of Neural Networks
Jing Wang
Bing Li
Grace Li Zhang
11
4
0
23 Sep 2023
Learning Accurate Performance Predictors for Ultrafast Automated Model
  Compression
Learning Accurate Performance Predictors for Ultrafast Automated Model Compression
Ziwei Wang
Jiwen Lu
Han Xiao
Shengyu Liu
Jie Zhou
OffRL
15
1
0
13 Apr 2023
Class-based Quantization for Neural Networks
Class-based Quantization for Neural Networks
Wenhao Sun
Grace Li Zhang
Huaxi Gu
Bing Li
Ulf Schlichtmann
MQ
19
7
0
27 Nov 2022
Collaborative Multi-Teacher Knowledge Distillation for Learning Low
  Bit-width Deep Neural Networks
Collaborative Multi-Teacher Knowledge Distillation for Learning Low Bit-width Deep Neural Networks
Cuong Pham
Tuan Hoang
Thanh-Toan Do
FedML
MQ
21
14
0
27 Oct 2022
DRESS: Dynamic REal-time Sparse Subnets
DRESS: Dynamic REal-time Sparse Subnets
Zhongnan Qu
Syed Shakib Sarwar
Xin Dong
Yuecheng Li
Huseyin Ekin Sumbul
B. D. Salvo
3DH
13
1
0
01 Jul 2022
A Comprehensive Survey on Model Quantization for Deep Neural Networks in
  Image Classification
A Comprehensive Survey on Model Quantization for Deep Neural Networks in Image Classification
Babak Rokh
A. Azarpeyvand
Alireza Khanteymoori
MQ
30
82
0
14 May 2022
Arrhythmia Classifier Using Convolutional Neural Network with Adaptive
  Loss-aware Multi-bit Networks Quantization
Arrhythmia Classifier Using Convolutional Neural Network with Adaptive Loss-aware Multi-bit Networks Quantization
Hanshi Sun
Ao Wang
Ninghao Pu
Zhiqing Li
Jung Y. Huang
Hao Liu
Zhiyu Qi
MQ
14
4
0
27 Feb 2022
Finding the Task-Optimal Low-Bit Sub-Distribution in Deep Neural
  Networks
Finding the Task-Optimal Low-Bit Sub-Distribution in Deep Neural Networks
Runpei Dong
Zhanhong Tan
Mengdi Wu
Linfeng Zhang
Kaisheng Ma
MQ
33
11
0
30 Dec 2021
A Generalized Zero-Shot Quantization of Deep Convolutional Neural
  Networks via Learned Weights Statistics
A Generalized Zero-Shot Quantization of Deep Convolutional Neural Networks via Learned Weights Statistics
Prasen Kumar Sharma
Arun Abraham
V. N. Rajendiran
MQ
25
7
0
06 Dec 2021
Sharpness-aware Quantization for Deep Neural Networks
Sharpness-aware Quantization for Deep Neural Networks
Jing Liu
Jianfei Cai
Bohan Zhuang
MQ
16
24
0
24 Nov 2021
Automatic Neural Network Pruning that Efficiently Preserves the Model
  Accuracy
Automatic Neural Network Pruning that Efficiently Preserves the Model Accuracy
Thibault Castells
Seul-Ki Yeom
3DV
16
3
0
18 Nov 2021
Generalizable Mixed-Precision Quantization via Attribution Rank
  Preservation
Generalizable Mixed-Precision Quantization via Attribution Rank Preservation
Ziwei Wang
Han Xiao
Jiwen Lu
Jie Zhou
MQ
14
32
0
05 Aug 2021
Training Compact CNNs for Image Classification using Dynamic-coded
  Filter Fusion
Training Compact CNNs for Image Classification using Dynamic-coded Filter Fusion
Mingbao Lin
Bohong Chen
Fei Chao
Rongrong Ji
VLM
22
23
0
14 Jul 2021
Model compression as constrained optimization, with application to
  neural nets. Part V: combining compressions
Model compression as constrained optimization, with application to neural nets. Part V: combining compressions
Miguel Á. Carreira-Perpiñán
Yerlan Idelbayev
17
6
0
09 Jul 2021
Measuring what Really Matters: Optimizing Neural Networks for TinyML
Measuring what Really Matters: Optimizing Neural Networks for TinyML
Lennart Heim
Andreas Biri
Zhongnan Qu
Lothar Thiele
38
30
0
21 Apr 2021
FTBNN: Rethinking Non-linearity for 1-bit CNNs and Going Beyond
FTBNN: Rethinking Non-linearity for 1-bit CNNs and Going Beyond
Z. Su
Linpu Fang
Deke Guo
Duwen Hu
M. Pietikäinen
Li Liu
MQ
11
3
0
19 Oct 2020
Deep Partial Updating: Towards Communication Efficient Updating for
  On-device Inference
Deep Partial Updating: Towards Communication Efficient Updating for On-device Inference
Zhongnan Qu
Cong Liu
Lothar Thiele
3DH
13
3
0
06 Jul 2020
Incremental Network Quantization: Towards Lossless CNNs with
  Low-Precision Weights
Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights
Aojun Zhou
Anbang Yao
Yiwen Guo
Lin Xu
Yurong Chen
MQ
311
1,047
0
10 Feb 2017
1