ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2308.11518
11
1

EM for Mixture of Linear Regression with Clustered Data

22 August 2023
Amirhossein Reisizadeh
Khashayar Gatmiry
Asuman Ozdaglar
    FedML
ArXivPDFHTML
Abstract

Modern data-driven and distributed learning frameworks deal with diverse massive data generated by clients spread across heterogeneous environments. Indeed, data heterogeneity is a major bottleneck in scaling up many distributed learning paradigms. In many settings however, heterogeneous data may be generated in clusters with shared structures, as is the case in several applications such as federated learning where a common latent variable governs the distribution of all the samples generated by a client. It is therefore natural to ask how the underlying clustered structures in distributed data can be exploited to improve learning schemes. In this paper, we tackle this question in the special case of estimating ddd-dimensional parameters of a two-component mixture of linear regressions problem where each of mmm nodes generates nnn samples with a shared latent variable. We employ the well-known Expectation-Maximization (EM) method to estimate the maximum likelihood parameters from mmm batches of dependent samples each containing nnn measurements. Discarding the clustered structure in the mixture model, EM is known to require O(log⁡(mn/d))O(\log(mn/d))O(log(mn/d)) iterations to reach the statistical accuracy of O(d/(mn))O(\sqrt{d/(mn)})O(d/(mn)​). In contrast, we show that if initialized properly, EM on the structured data requires only O(1)O(1)O(1) iterations to reach the same statistical accuracy, as long as mmm grows up as eo(n)e^{o(n)}eo(n). Our analysis establishes and combines novel asymptotic optimization and generalization guarantees for population and empirical EM with dependent samples, which may be of independent interest.

View on arXiv
Comments on this paper