ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2401.15347
  4. Cited By
A Comprehensive Survey of Compression Algorithms for Language Models

A Comprehensive Survey of Compression Algorithms for Language Models

27 January 2024
Seungcheol Park
Jaehyeon Choi
Sojin Lee
U. Kang
    MQ
ArXiv (abs)PDFHTML

Papers citing "A Comprehensive Survey of Compression Algorithms for Language Models"

11 / 11 papers shown
Global Cross-Time Attention Fusion for Enhanced Solar Flare Prediction from Multivariate Time Series
Global Cross-Time Attention Fusion for Enhanced Solar Flare Prediction from Multivariate Time Series
Onur Vural
S. M. Hamdi
S. F. Boubrahimi
AI4TS
139
0
0
17 Nov 2025
Unifying Uniform and Binary-coding Quantization for Accurate Compression of Large Language Models
Unifying Uniform and Binary-coding Quantization for Accurate Compression of Large Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Seungcheol Park
Jeongin Bae
Beomseok Kwon
Minjun Kim
Byeongwook Kim
S. Kwon
U. Kang
Dongsoo Lee
MQ
379
0
0
04 Jun 2025
Accurate Sublayer Pruning for Large Language Models by Exploiting Latency and Tunability Information
Accurate Sublayer Pruning for Large Language Models by Exploiting Latency and Tunability InformationInternational Joint Conference on Artificial Intelligence (IJCAI), 2025
Seungcheol Park
Sojin Lee
Jongjin Kim
Jinsik Lee
Hyunjik Jo
U. Kang
276
3
0
04 Jun 2025
QA-HFL: Quality-Aware Hierarchical Federated Learning for Resource-Constrained Mobile Devices with Heterogeneous Image Quality
QA-HFL: Quality-Aware Hierarchical Federated Learning for Resource-Constrained Mobile Devices with Heterogeneous Image Quality
Sajid Hussain
Muhammad Sohail
Nauman Ali Khan
379
4
0
04 Jun 2025
Zero-shot Quantization: A Comprehensive Survey
Zero-shot Quantization: A Comprehensive SurveyInternational Joint Conference on Artificial Intelligence (IJCAI), 2024
Minjun Kim
Jaehyeon Choi
Jongkeun Lee
Wonjin Cho
U. Kang
MQ
369
6
0
14 May 2025
Benchmarking Post-Training Quantization in LLMs: Comprehensive Taxonomy, Unified Evaluation, and Comparative Analysis
Benchmarking Post-Training Quantization in LLMs: Comprehensive Taxonomy, Unified Evaluation, and Comparative Analysis
Jiaqi Zhao
Ming Wang
Miao Zhang
Yuzhang Shang
Xuebo Liu
Yaowei Wang
Min Zhang
Liqiang Nie
MQ
633
6
0
18 Feb 2025
Large Language Model Inference Acceleration: A Comprehensive Hardware Perspective
Large Language Model Inference Acceleration: A Comprehensive Hardware Perspective
Jinhao Li
Jiaming Xu
Shan Huang
Yonghua Chen
Wen Li
...
Jiayi Pan
Li Ding
Hao Zhou
Yu Wang
Guohao Dai
633
49
0
06 Oct 2024
The Early Bird Catches the Leak: Unveiling Timing Side Channels in LLM Serving Systems
The Early Bird Catches the Leak: Unveiling Timing Side Channels in LLM Serving SystemsIEEE Transactions on Information Forensics and Security (IEEE TIFS), 2024
Linke Song
Zixuan Pang
Wenhao Wang
Zihao Wang
XiaoFeng Wang
H. G. Chen
Wei Song
Yier Jin
Dan Meng
Rui Hou
617
18
0
30 Sep 2024
Compact Language Models via Pruning and Knowledge Distillation
Compact Language Models via Pruning and Knowledge Distillation
Saurav Muralidharan
Sharath Turuvekere Sreenivas
Raviraj Joshi
Marcin Chochowski
M. Patwary
Mohammad Shoeybi
Bryan Catanzaro
Jan Kautz
Pavlo Molchanov
SyDaMQ
357
117
0
19 Jul 2024
Accurate Retraining-free Pruning for Pretrained Encoder-based Language
  Models
Accurate Retraining-free Pruning for Pretrained Encoder-based Language ModelsInternational Conference on Learning Representations (ICLR), 2023
Seungcheol Park
Ho-Jin Choi
U. Kang
VLM
237
12
0
07 Aug 2023
Fast Transformer Decoding: One Write-Head is All You Need
Fast Transformer Decoding: One Write-Head is All You Need
Noam M. Shazeer
599
641
0
06 Nov 2019
1
Page 1 of 1