ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.00572
31
0

LoR2C : Low-Rank Residual Connection Adaptation for Parameter-Efficient Fine-Tuning

1 March 2025
Jiancheng Zhao
Xingda Yu
Yuxiang Zhang
Zhen Yang
    OffRL
ArXivPDFHTML
Abstract

In recent years, pretrained large language models have demonstrated outstanding performance across various natural language processing tasks. However, full-parameter fine-tuning methods require adjusting all model parameters, leading to immense computational resource demands. Although parameter-efficient fine-tuning methods like LoRA have significantly reduced the number of parameters, they still face challenges such as gradient vanishing and the potential for further parameter reduction. To address these issues, this paper proposes a novel parameter-efficient fine-tuning method called LoR2C (Low-Rank Residual Connection Adaptation). LoR2C introduces residual connections with low-rank matrices within the model layers, which not only reduces the number of fine-tuning parameters but also effectively alleviates the gradient vanishing problem. Additionally, this paper presents three optimization variants of LoR2C: ShareLoR2C, MergeLoR2C, and InjectLoR2C. These variants further improve parameter efficiency and model performance through parameter sharing, module merging, and injection mechanisms, respectively. Experimental results on multiple natural language understanding and natural language generation tasks demonstrate that LoR2C and its optimized variants significantly reduce parameter overhead while maintaining or even improving performance, outperforming existing mainstream parameter-efficient fine-tuningthis http URLcode is publicly available atthis https URL.

View on arXiv
@article{zhao2025_2503.00572,
  title={ LoR2C : Low-Rank Residual Connection Adaptation for Parameter-Efficient Fine-Tuning },
  author={ Jiancheng Zhao and Xingda Yu and Yuxiang Zhang and Zhen Yang },
  journal={arXiv preprint arXiv:2503.00572},
  year={ 2025 }
}
Comments on this paper