ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2003.09603
21
66

Dynamic Sampling and Selective Masking for Communication-Efficient Federated Learning

21 March 2020
Shaoxiong Ji
Wenqi Jiang
A. Walid
Xue Li
    FedML
ArXivPDFHTML
Abstract

Federated learning (FL) is a novel machine learning setting that enables on-device intelligence via decentralized training and federated optimization. Deep neural networks' rapid development facilitates the learning techniques for modeling complex problems and emerges into federated deep learning under the federated setting. However, the tremendous amount of model parameters burdens the communication network with a high load of transportation. This paper introduces two approaches for improving communication efficiency by dynamic sampling and top-kkk selective masking. The former controls the fraction of selected client models dynamically, while the latter selects parameters with top-kkk largest values of difference for federated updating. Experiments on convolutional image classification and recurrent language modeling are conducted on three public datasets to show our proposed methods' effectiveness.

View on arXiv
Comments on this paper