ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.10412
51
0

dFLMoE: Decentralized Federated Learning via Mixture of Experts for Medical Data Analysis

13 March 2025
Luyuan Xie
Tianyu Luan
Wenyuan Cai
Guochen Yan
Z. Chen
Nan Xi
Yuejian Fang
Qingni Shen
Zhonghai Wu
Junsong Yuan
    FedML
ArXivPDFHTML
Abstract

Federated learning has wide applications in the medical field. It enables knowledge sharing among different healthcare institutes while protecting patients' privacy. However, existing federated learning systems are typically centralized, requiring clients to upload client-specific knowledge to a central server for aggregation. This centralized approach would integrate the knowledge from each client into a centralized server, and the knowledge would be already undermined during the centralized integration before it reaches back to each client. Besides, the centralized approach also creates a dependency on the central server, which may affect training stability if the server malfunctions or connections are unstable. To address these issues, we propose a decentralized federated learning framework named dFLMoE. In our framework, clients directly exchange lightweight head models with each other. After exchanging, each client treats both local and received head models as individual experts, and utilizes a client-specific Mixture of Experts (MoE) approach to make collective decisions. This design not only reduces the knowledge damage with client-specific aggregations but also removes the dependency on the central server to enhance the robustness of the framework. We validate our framework on multiple medical tasks, demonstrating that our method evidently outperforms state-of-the-art approaches under both model homogeneity and heterogeneity settings.

View on arXiv
@article{xie2025_2503.10412,
  title={ dFLMoE: Decentralized Federated Learning via Mixture of Experts for Medical Data Analysis },
  author={ Luyuan Xie and Tianyu Luan and Wenyuan Cai and Guochen Yan and Zhaoyu Chen and Nan Xi and Yuejian Fang and Qingni Shen and Zhonghai Wu and Junsong Yuan },
  journal={arXiv preprint arXiv:2503.10412},
  year={ 2025 }
}
Comments on this paper