59
0

WeightLoRA: Keep Only Necessary Adapters

Main:8 Pages
3 Figures
Bibliography:3 Pages
11 Tables
Appendix:2 Pages
Abstract

The widespread utilization of language models in modern applications is inconceivable without Parameter-Efficient Fine-Tuning techniques, such as low-rank adaptation (LoRA\texttt{LoRA}), which adds trainable adapters to selected layers. Although LoRA\texttt{LoRA} may obtain accurate solutions, it requires significant memory to train large models and intuition on which layers to add adapters. In this paper, we propose a novel method, WeightLoRA\texttt{WeightLoRA}, which overcomes this issue by adaptive selection of the most critical LoRA\texttt{LoRA} heads throughout the optimization process. As a result, we can significantly reduce the number of trainable parameters while maintaining the capability to obtain consistent or even superior metric values. We conduct experiments for a series of competitive benchmarks and DeBERTa, BART, and Llama models, comparing our method with different adaptive approaches. The experimental results demonstrate the efficacy of WeightLoRA\texttt{WeightLoRA} and the superior performance of WeightLoRA+\texttt{WeightLoRA+} in almost all cases.

View on arXiv
@article{veprikov2025_2506.02724,
  title={ WeightLoRA: Keep Only Necessary Adapters },
  author={ Andrey Veprikov and Vladimir Solodkin and Alexander Zyl and Andrey Savchenko and Aleksandr Beznosikov },
  journal={arXiv preprint arXiv:2506.02724},
  year={ 2025 }
}
Comments on this paper