ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2309.14021
  4. Cited By
LORD: Low Rank Decomposition Of Monolingual Code LLMs For One-Shot
  Compression

LORD: Low Rank Decomposition Of Monolingual Code LLMs For One-Shot Compression

25 September 2023
Ayush Kaushal
Tejas Vaidhya
Irina Rish
ArXiv (abs)PDFHTMLHuggingFace (1 upvotes)Github (430★)

Papers citing "LORD: Low Rank Decomposition Of Monolingual Code LLMs For One-Shot Compression"

14 / 14 papers shown
FLRC: Fine-grained Low-Rank Compressor for Efficient LLM Inference
FLRC: Fine-grained Low-Rank Compressor for Efficient LLM Inference
Yu-Chen Lu
Chong-Yan Chen
Chi-Chih Chang
Yu-Fang Hu
Kai-Chiang Wu
116
1
0
10 Oct 2025
Accelerating Attention with Basis Decomposition
Accelerating Attention with Basis Decomposition
Jialin Zhao
193
0
0
02 Oct 2025
CALR: Corrective Adaptive Low-Rank Decomposition for Efficient Large Language Model Layer Compression
CALR: Corrective Adaptive Low-Rank Decomposition for Efficient Large Language Model Layer Compression
Muchammad Daniyal Kautsar
Afra Majida Hariono
Widyawan
Syukron Abu Ishaq Alfarozi
Kuntpong Woraratpanya
197
0
0
21 Aug 2025
Efficient Large Language Model Inference with Neural Block Linearization
Efficient Large Language Model Inference with Neural Block Linearization
Mete Erdogan
F. Tonin
Volkan Cevher
430
1
0
27 May 2025
ResSVD: Residual Compensated SVD for Large Language Model Compression
ResSVD: Residual Compensated SVD for Large Language Model Compression
Haolei Bai
Siyong Jian
Tuo Liang
Yu Yin
Huan Wang
410
5
0
26 May 2025
Lillama: Large Language Models Compression via Low-Rank Feature Distillation
Lillama: Large Language Models Compression via Low-Rank Feature Distillation
Yaya Sy
Christophe Cerisara
Irina Illina
MQ
345
0
0
31 Dec 2024
Beware of Calibration Data for Pruning Large Language Models
Beware of Calibration Data for Pruning Large Language ModelsInternational Conference on Learning Representations (ICLR), 2024
Yixin Ji
Yang Xiang
Juntao Li
Qingrong Xia
Ping Li
Xinyu Duan
Zhefeng Wang
Min Zhang
351
9
0
23 Oct 2024
Eigen Attention: Attention in Low-Rank Space for KV Cache Compression
Eigen Attention: Attention in Low-Rank Space for KV Cache CompressionConference on Empirical Methods in Natural Language Processing (EMNLP), 2024
Utkarsh Saxena
Gobinda Saha
Sakshi Choudhary
Kaushik Roy
303
44
0
10 Aug 2024
A Survey on Efficient Inference for Large Language Models
A Survey on Efficient Inference for Large Language Models
Zixuan Zhou
Xuefei Ning
Ke Hong
Tianyu Fu
Jiaming Xu
...
Shengen Yan
Guohao Dai
Xiao-Ping Zhang
Yuhan Dong
Yu Wang
481
205
0
22 Apr 2024
LoRAP: Transformer Sub-Layers Deserve Differentiated Structured
  Compression for Large Language Models
LoRAP: Transformer Sub-Layers Deserve Differentiated Structured Compression for Large Language Models
Guangyan Li
Yongqiang Tang
Wensheng Zhang
317
9
0
15 Apr 2024
Model Compression and Efficient Inference for Large Language Models: A
  Survey
Model Compression and Efficient Inference for Large Language Models: A Survey
Wenxiao Wang
Wei Chen
Yicong Luo
Yongliu Long
Zhengkai Lin
Liye Zhang
Binbin Lin
Deng Cai
Xiaofei He
MQ
362
93
0
15 Feb 2024
A Survey on Transformer Compression
A Survey on Transformer Compression
Yehui Tang
Yunhe Wang
Jianyuan Guo
Zhijun Tu
Kai Han
Hailin Hu
Dacheng Tao
563
74
0
05 Feb 2024
Faster and Lighter LLMs: A Survey on Current Challenges and Way Forward
Faster and Lighter LLMs: A Survey on Current Challenges and Way Forward
Arnav Chavan
Raghav Magazine
Shubham Kushwaha
M. Debbah
Deepak Gupta
362
41
0
02 Feb 2024
Fast Transformer Decoding: One Write-Head is All You Need
Fast Transformer Decoding: One Write-Head is All You Need
Noam M. Shazeer
808
716
0
06 Nov 2019
1
Page 1 of 1