Low-Rank Matrix Approximation for Neural Network Compression

Deep Neural Networks (DNNs) have encountered an emerging deployment challenge due to large and expensive memory and computation requirements. In this paper, we present a new Adaptive-Rank Singular Value Decomposition (ARSVD) method that approximates the optimal rank for compressing weight matrices in neural networks using spectral entropy. Unlike conventional SVD-based methods that apply a fixed-rank truncation across all layers, ARSVD uses an adaptive selection of the rank per layer through the entropy distribution of its singular values. This approach ensures that each layer will retain a certain amount of its informational content, thereby reducing redundancy. Our method enables efficient, layer-wise compression, yielding improved performance with reduced space and time complexity compared to static-rank reduction techniques.
View on arXiv@article{cherukuri2025_2504.20078, title={ Low-Rank Matrix Approximation for Neural Network Compression }, author={ Kalyan Cherukuri and Aarav Lala }, journal={arXiv preprint arXiv:2504.20078}, year={ 2025 } }