26
0

Federated Adapter on Foundation Models: An Out-Of-Distribution Approach

Abstract

As foundation models gain prominence, Federated Foundation Models (FedFM) have emerged as a privacy-preserving approach to collaboratively fine-tune models in federated learning (FL) frameworks using distributed datasets across clients. A key challenge for FedFM, given the versatile nature of foundation models, is addressing out-of-distribution (OOD) generalization, where unseen tasks or clients may exhibit distribution shifts leading to suboptimal performance. Although numerous studies have explored OOD generalization in conventional FL, these methods are inadequate for FedFM due to the challenges posed by large parameter scales and increased data heterogeneity. To address these, we propose FedOA, which employs adapter-based parameter-efficient fine-tuning methods for efficacy and introduces personalized adapters with feature distance-based regularization to align distributions and guarantee OOD generalization for each client. Theoretically, we demonstrate that the conventional aggregated global model in FedFM inherently retains OOD generalization capabilities, and our proposed method enhances the personalized model's OOD generalization through regularization informed by the global model, with proven convergence under general non-convex settings. Empirically, the effectiveness of the proposed method is validated on benchmark datasets across various NLP tasks.

View on arXiv
@article{yang2025_2505.01075,
  title={ Federated Adapter on Foundation Models: An Out-Of-Distribution Approach },
  author={ Yiyuan Yang and Guodong Long and Tianyi Zhou and Qinghua Lu and Shanshan Ye and Jing Jiang },
  journal={arXiv preprint arXiv:2505.01075},
  year={ 2025 }
}
Comments on this paper