-Tuning: An Efficient Tuning Paradigm for Large-Scale Pre-Trained Models via Label Representation Learning

With the success of large-scale pre-trained models (PTMs), how efficiently adapting PTMs to downstream tasks has attracted tremendous attention, especially for PTMs with billions of parameters. Although some parameter-efficient tuning paradigms have been proposed to address this problem, they still require large resources to compute the gradients in the training phase. In this paper, we propose -Tuning, an efficient yet effective paradigm to adapt frozen large-scale PTMs to specific downstream tasks. -tuning learns dense representations for labels defined in a given task and aligns them to fixed feature representation. Without tuning the features of input text and model parameters, -tuning is both parameter-efficient and training-efficient. For with 1.6 billion parameters, -tuning achieves performance more than of full fine-tuning on GLUE Benchmark with only tunable parameters and much fewer training costs.
View on arXiv