Block Circulant Adapter for Large Language Models

Abstract
Fine-tuning large language models (LLMs) is difficult due to their huge model size. Recent Fourier domain-based methods show potential for reducing fine-tuning costs. We propose a block circulant matrix-based fine-tuning method with a stable training heuristic to leverage the properties of circulant matrices and one-dimensional Fourier transforms to reduce storage and computation costs. Experiments show that our method uses less number of parameters than VeRA, smaller than LoRA and less FLOPs than FourierFT, while maintaining close or better task performance. Our approach presents a promising way in frequency domain to fine-tune large models on downstream tasks.
View on arXiv@article{ding2025_2505.00582, title={ Block Circulant Adapter for Large Language Models }, author={ Xinyu Ding and Meiqi Wang and Siyu Liao and Zhongfeng Wang }, journal={arXiv preprint arXiv:2505.00582}, year={ 2025 } }
Comments on this paper