26
71

Parameter-Efficient Fine-Tuning for Pre-Trained Vision Models: A Survey

Yi Xin
Jianjiang Yang
Haodi Zhou
Junlong Du
Junlong Du
Yue Fan
Qing Li
Qing Li
Yuntao Du
Abstract

Large-scale pre-trained vision models (PVMs) have shown great potential for adaptability across various downstream vision tasks. However, with state-of-the-art PVMs growing to billions or even trillions of parameters, the standard full fine-tuning paradigm is becoming unsustainable due to high computational and storage demands. In response, researchers are exploring parameter-efficient fine-tuning (PEFT), which seeks to exceed the performance of full fine-tuning with minimal parameter modifications. This survey provides a comprehensive overview and future directions for visual PEFT, offering a systematic review of the latest advancements. First, we provide a formal definition of PEFT and discuss model pre-training methods. We then categorize existing methods into three categories: addition-based, partial-based, and unified-based. Finally, we introduce the commonly used datasets and applications and suggest potential future research challenges. A comprehensive collection of resources is available atthis https URL.

View on arXiv
@article{xin2025_2402.02242,
  title={ Parameter-Efficient Fine-Tuning for Pre-Trained Vision Models: A Survey },
  author={ Yi Xin and Jianjiang Yang and Siqi Luo and Haodi Zhou and Junlong Du and Xiaohong Liu and Yue Fan and Qing Li and Yuntao Du },
  journal={arXiv preprint arXiv:2402.02242},
  year={ 2025 }
}
Comments on this paper