ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2210.06384
  4. Cited By
GMP*: Well-Tuned Gradual Magnitude Pruning Can Outperform Most
  BERT-Pruning Methods
v1v2v3 (latest)

GMP*: Well-Tuned Gradual Magnitude Pruning Can Outperform Most BERT-Pruning Methods

12 October 2022
Eldar Kurtic
Dan Alistarh
    AI4MH
ArXiv (abs)PDFHTMLHuggingFace (1 upvotes)Github (159486★)

Papers citing "GMP*: Well-Tuned Gradual Magnitude Pruning Can Outperform Most BERT-Pruning Methods"

15 / 15 papers shown
CAST: Continuous and Differentiable Semi-Structured Sparsity-Aware Training for Large Language Models
CAST: Continuous and Differentiable Semi-Structured Sparsity-Aware Training for Large Language Models
Weiyu Huang
Yuezhou Hu
Jun Zhu
Jianfei Chen
CLL
142
0
0
30 Sep 2025
SAFE: Finding Sparse and Flat Minima to Improve Pruning
SAFE: Finding Sparse and Flat Minima to Improve Pruning
Dongyeop Lee
Kwanhee Lee
Jinseok Chung
Namhoon Lee
394
5
0
07 Jun 2025
Pruning Large Language Models with Semi-Structural Adaptive Sparse
  Training
Pruning Large Language Models with Semi-Structural Adaptive Sparse Training
Weiyu Huang
Yuezhou Hu
Guohao Jian
Jun Zhu
Jianfei Chen
384
24
0
30 Jul 2024
Beyond Perplexity: Multi-dimensional Safety Evaluation of LLM
  Compression
Beyond Perplexity: Multi-dimensional Safety Evaluation of LLM Compression
Zhichao Xu
Ashim Gupta
Tao Li
Oliver Bentham
Vivek Srikumar
499
29
0
06 Jul 2024
SPD-CFL: Stepwise Parameter Dropout for Efficient Continual Federated Learning
SPD-CFL: Stepwise Parameter Dropout for Efficient Continual Federated Learning
Yuning Yang
H. Yu
Tianrun Gao
Tianrun Gao
Xiaohong Liu
Xiaodong Xu
Ping Zhang
Guangyu Wang
352
6
0
15 May 2024
Model Compression and Efficient Inference for Large Language Models: A
  Survey
Model Compression and Efficient Inference for Large Language Models: A Survey
Wenxiao Wang
Wei Chen
Yicong Luo
Yongliu Long
Zhengkai Lin
Liye Zhang
Binbin Lin
Deng Cai
Xiaofei He
MQ
379
95
0
15 Feb 2024
VENOM: A Vectorized N:M Format for Unleashing the Power of Sparse Tensor
  Cores
VENOM: A Vectorized N:M Format for Unleashing the Power of Sparse Tensor CoresInternational Conference for High Performance Computing, Networking, Storage and Analysis (SC), 2023
Roberto L. Castro
Andrei Ivanov
Diego Andrade
Tal Ben-Nun
B. Fraguela
Torsten Hoefler
247
35
0
03 Oct 2023
Scaling Laws for Sparsely-Connected Foundation Models
Scaling Laws for Sparsely-Connected Foundation ModelsInternational Conference on Learning Representations (ICLR), 2023
Elias Frantar
C. Riquelme
N. Houlsby
Dan Alistarh
Utku Evci
386
46
0
15 Sep 2023
How Does Pruning Impact Long-Tailed Multi-Label Medical Image
  Classifiers?
How Does Pruning Impact Long-Tailed Multi-Label Medical Image Classifiers?International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), 2023
G. Holste
Ziyu Jiang
Ajay Jaiswal
Maria Hanna
Shlomo Minkowitz
...
Ying Ding
Ronald M. Summers
George Shih
Yifan Peng
Zhangyang Wang
302
1
0
17 Aug 2023
Accurate Neural Network Pruning Requires Rethinking Sparse Optimization
Accurate Neural Network Pruning Requires Rethinking Sparse Optimization
Denis Kuznedelev
Eldar Kurtic
Eugenia Iofinova
Elias Frantar
Alexandra Peste
Dan Alistarh
VLM
381
13
0
03 Aug 2023
Towards Automated Circuit Discovery for Mechanistic Interpretability
Towards Automated Circuit Discovery for Mechanistic InterpretabilityNeural Information Processing Systems (NeurIPS), 2023
Arthur Conmy
Augustine N. Mavor-Parker
Aengus Lynch
Stefan Heimersheim
Adrià Garriga-Alonso
636
539
0
28 Apr 2023
Bias in Pruned Vision Models: In-Depth Analysis and Countermeasures
Bias in Pruned Vision Models: In-Depth Analysis and CountermeasuresComputer Vision and Pattern Recognition (CVPR), 2023
Eugenia Iofinova
Alexandra Peste
Dan Alistarh
306
14
0
25 Apr 2023
oBERTa: Improving Sparse Transfer Learning via improved initialization,
  distillation, and pruning regimes
oBERTa: Improving Sparse Transfer Learning via improved initialization, distillation, and pruning regimes
Daniel Fernando Campos
Alexandre Marques
Mark Kurtz
Chengxiang Zhai
VLMAAML
270
4
0
30 Mar 2023
ZipLM: Inference-Aware Structured Pruning of Language Models
ZipLM: Inference-Aware Structured Pruning of Language ModelsNeural Information Processing Systems (NeurIPS), 2023
Eldar Kurtic
Elias Frantar
Dan Alistarh
MQ
455
52
0
07 Feb 2023
SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot
SparseGPT: Massive Language Models Can Be Accurately Pruned in One-ShotInternational Conference on Machine Learning (ICML), 2023
Elias Frantar
Dan Alistarh
VLM
773
1,199
0
02 Jan 2023
1
Page 1 of 1