ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.15092
58
0

Optimizing Singular Spectrum for Large Language Model Compression

24 February 2025
Dengjie Li
Tiancheng Shen
Yao Zhou
Baisong Yang
Zhongying Liu
Masheng Yang
Bernard Ghanem
Yibo Yang
Yujie Zhong
Ming-Hsuan Yang
ArXivPDFHTML
Abstract

Large language models (LLMs) have demonstrated remarkable capabilities, yet prohibitive parameter complexity often hinders their deployment. Existing singular value decomposition (SVD) based compression methods simply deem singular values as importance scores of decomposed components. However, this importance ordered by singular values does not necessarily correlate with the performance of a downstream task. In this work, we introduce SoCo (Singular spectrum optimization for large language model Compression), a novel compression framework that learns to rescale the decomposed components of SVD in a data-driven manner. Concretely, we employ a learnable diagonal matrix to assign importance scores for singular spectrum and develop a three-stage training process that progressively refines these scores from initial coarse compression to fine-grained sparsification-thereby striking an effective balance between aggressive model compression and performance preservation. Thanks to the learnable singular spectrum, SoCo adaptively prunes components according to the sparsified importance scores, rather than relying on the fixed order of singular values. More importantly, the remaining components with amplified importance scores can compensate for the loss of the pruned ones. Experimental evaluations across multiple LLMs and benchmarks demonstrate that SoCo surpasses the state-of-the-art methods in model compression.

View on arXiv
@article{li2025_2502.15092,
  title={ Optimizing Singular Spectrum for Large Language Model Compression },
  author={ Dengjie Li and Tiancheng Shen and Yao Zhou and Baisong Yang and Zhongying Liu and Masheng Yang and Bernard Ghanem and Yibo Yang and Yujie Zhong and Ming-Hsuan Yang },
  journal={arXiv preprint arXiv:2502.15092},
  year={ 2025 }
}
Comments on this paper