Federated Learning (FL) enables distributed ML model training on private user data at the global scale. Despite the potential of FL demonstrated in many domains, an in-depth view of its impact on model accuracy remains unclear. In this paper, we investigate, systematically, how this learning paradigm can affect the accuracy of state-of-the-art ML models for a variety of ML tasks. We present an empirical study that involves various data types: text, image, audio, and video, and FL configuration knobs: data distribution, FL scale, client sampling, and local and global computations. Our experiments are conducted in a unified FL framework to achieve high fidelity, with substantial human efforts and resource investments. Based on the results, we perform a quantitative analysis of the impact of FL, and highlight challenging scenarios where applying FL degrades the accuracy of the model drastically and identify cases where the impact is negligible. The detailed and extensive findings can benefit practical deployments and future development of FL.
View on arXiv@article{yang2025_2503.20768, title={ An Empirical Study of the Impact of Federated Learning on Machine Learning Model Accuracy }, author={ Haotian Yang and Zhuoran Wang and Benson Chou and Sophie Xu and Hao Wang and Jingxian Wang and Qizhen Zhang }, journal={arXiv preprint arXiv:2503.20768}, year={ 2025 } }