147

LoRALib: A Standardized Benchmark for Evaluating LoRA-MoE Methods

Main:4 Pages
1 Figures
Bibliography:1 Pages
3 Tables
Abstract

As a parameter efficient fine-tuning (PEFT) method, low-rank adaptation (LoRA) can save significant costs in storage and computing, but its strong adaptability to a single task is often accompanied by insufficient cross-task generalization capabilities. To improve this, existing work combines LoRA with mixture-of-experts (MoE) to enhance the model's adaptability through expert modules and routing mechanisms. However, existing LoRA-MoE methods lack unified standards in models, datasets, hyperparameters, and evaluation methods, making it difficult to conduct fair comparisons between different methods. To this end, we proposed a unified benchmark named LoRALib. Specifically, we standardized datasets from 4040 downstream tasks into a unified format, fine-tuned them using the same hyperparameters and obtained 680680 LoRA modules across 1717 model architectures. Based on this LoRA library, we conduct large-scale experiments on 33 representative LoRA-MoE methods and different LoRA selection mechanisms using the open-sourced testing tool OpenCompass. Extensive experiments show that LoRAMoE performs best, and that prioritizing LoRAs relevant to the target task can further improve the performance of MoE. We hope these findings will inspire future work. Our datasets and LoRA library are available atthis https URLandthis https URL.

View on arXiv
Comments on this paper