36
0

CE-LoRA: Computation-Efficient LoRA Fine-Tuning for Language Models

Abstract

Large Language Models (LLMs) demonstrate exceptional performance across various tasks but demand substantial computational resources even for fine-tuning computation. Although Low-Rank Adaptation (LoRA) significantly alleviates memory consumption during fine-tuning, its impact on computational cost reduction is limited. This paper identifies the computation of activation gradients as the primary bottleneck in LoRA's backward propagation and introduces the Computation-Efficient LoRA (CE-LoRA) algorithm, which enhances computational efficiency while preserving memory efficiency. CE-LoRA leverages two key techniques: Approximated Matrix Multiplication, which replaces dense multiplications of large and complete matrices with sparse multiplications involving only critical rows and columns, and the Double-LoRA technique, which reduces error propagation in activation gradients. Theoretically, CE-LoRA converges at the same rate as LoRA, O(1/T) \mathcal{O}(1/\sqrt{T}) , where TT is the number of iteartions. Empirical evaluations confirm that CE-LoRA significantly reduces computational costs compared to LoRA without notable performance degradation.

View on arXiv
@article{chen2025_2502.01378,
  title={ CE-LoRA: Computation-Efficient LoRA Fine-Tuning for Language Models },
  author={ Guanduo Chen and Yutong He and Yipeng Hu and Kun Yuan and Binhang Yuan },
  journal={arXiv preprint arXiv:2502.01378},
  year={ 2025 }
}
Comments on this paper