ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2012.07774
17
32

Small Covers for Near-Zero Sets of Polynomials and Learning Latent Variable Models

14 December 2020
Ilias Diakonikolas
D. Kane
ArXivPDFHTML
Abstract

Let VVV be any vector space of multivariate degree-ddd homogeneous polynomials with co-dimension at most kkk, and SSS be the set of points where all polynomials in VVV {\em nearly} vanish. We establish a qualitatively optimal upper bound on the size of ϵ\epsilonϵ-covers for SSS, in the ℓ2\ell_2ℓ2​-norm. Roughly speaking, we show that there exists an ϵ\epsilonϵ-cover for SSS of cardinality M=(k/ϵ)Od(k1/d)M = (k/\epsilon)^{O_d(k^{1/d})}M=(k/ϵ)Od​(k1/d). Our result is constructive yielding an algorithm to compute such an ϵ\epsilonϵ-cover that runs in time poly(M)\mathrm{poly}(M)poly(M). Building on our structural result, we obtain significantly improved learning algorithms for several fundamental high-dimensional probabilistic models with hidden variables. These include density and parameter estimation for kkk-mixtures of spherical Gaussians (with known common covariance), PAC learning one-hidden-layer ReLU networks with kkk hidden units (under the Gaussian distribution), density and parameter estimation for kkk-mixtures of linear regressions (with Gaussian covariates), and parameter estimation for kkk-mixtures of hyperplanes. Our algorithms run in time {\em quasi-polynomial} in the parameter kkk. Previous algorithms for these problems had running times exponential in kΩ(1)k^{\Omega(1)}kΩ(1). At a high-level our algorithms for all these learning problems work as follows: By computing the low-degree moments of the hidden parameters, we are able to find a vector space of polynomials that nearly vanish on the unknown parameters. Our structural result allows us to compute a quasi-polynomial sized cover for the set of hidden parameters, which we exploit in our learning algorithms.

View on arXiv
Comments on this paper