24
0

Block Circulant Adapter for Large Language Models

Abstract

Fine-tuning large language models (LLMs) is difficult due to their huge model size. Recent Fourier domain-based methods show potential for reducing fine-tuning costs. We propose a block circulant matrix-based fine-tuning method with a stable training heuristic to leverage the properties of circulant matrices and one-dimensional Fourier transforms to reduce storage and computation costs. Experiments show that our method uses 14×14\times less number of parameters than VeRA, 16×16\times smaller than LoRA and 32×32\times less FLOPs than FourierFT, while maintaining close or better task performance. Our approach presents a promising way in frequency domain to fine-tune large models on downstream tasks.

View on arXiv
@article{ding2025_2505.00582,
  title={ Block Circulant Adapter for Large Language Models },
  author={ Xinyu Ding and Meiqi Wang and Siyu Liao and Zhongfeng Wang },
  journal={arXiv preprint arXiv:2505.00582},
  year={ 2025 }
}
Comments on this paper