ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2103.15653
200
8

The EM Algorithm is Adaptively-Optimal for Unbalanced Symmetric Gaussian Mixtures

Journal of machine learning research (JMLR), 2021
29 March 2021
Nir Weinberger
Guy Bresler
    FedML
ArXiv (abs)PDFHTML
Abstract

This paper studies the problem of estimating the means ±θ∗∈Rd\pm\theta_{*}\in\mathbb{R}^{d}±θ∗​∈Rd of a symmetric two-component Gaussian mixture δ∗⋅N(θ∗,I)+(1−δ∗)⋅N(−θ∗,I)\delta_{*}\cdot N(\theta_{*},I)+(1-\delta_{*})\cdot N(-\theta_{*},I)δ∗​⋅N(θ∗​,I)+(1−δ∗​)⋅N(−θ∗​,I) where the weights δ∗\delta_{*}δ∗​ and 1−δ∗1-\delta_{*}1−δ∗​ are unequal. Assuming that δ∗\delta_{*}δ∗​ is known, we show that the population version of the EM algorithm globally converges if the initial estimate has non-negative inner product with the mean of the larger weight component. This can be achieved by the trivial initialization θ0=0\theta_{0}=0θ0​=0. For the empirical iteration based on nnn samples, we show that when initialized at θ0=0\theta_{0}=0θ0​=0, the EM algorithm adaptively achieves the minimax error rate O~(min⁡{1(1−2δ∗)dn,1∥θ∗∥dn,(dn)1/4})\tilde{O}\Big(\min\Big\{\frac{1}{(1-2\delta_{*})}\sqrt{\frac{d}{n}},\frac{1}{\|\theta_{*}\|}\sqrt{\frac{d}{n}},\left(\frac{d}{n}\right)^{1/4}\Big\}\Big)O~(min{(1−2δ∗​)1​nd​​,∥θ∗​∥1​nd​​,(nd​)1/4}) in no more than O(1∥θ∗∥(1−2δ∗))O\Big(\frac{1}{\|\theta_{*}\|(1-2\delta_{*})}\Big)O(∥θ∗​∥(1−2δ∗​)1​) iterations (with high probability). We also consider the EM iteration for estimating the weight δ∗\delta_{*}δ∗​, assuming a fixed mean θ\thetaθ (which is possibly mismatched to θ∗\theta_{*}θ∗​). For the empirical iteration of nnn samples, we show that the minimax error rate O~(1∥θ∗∥dn)\tilde{O}\Big(\frac{1}{\|\theta_{*}\|}\sqrt{\frac{d}{n}}\Big)O~(∥θ∗​∥1​nd​​) is achieved in no more than O(1∥θ∗∥2)O\Big(\frac{1}{\|\theta_{*}\|^{2}}\Big)O(∥θ∗​∥21​) iterations. These results robustify and complement recent results of Wu and Zhou obtained for the equal weights case δ∗=1/2\delta_{*}=1/2δ∗​=1/2.

View on arXiv
Comments on this paper