92
3

A super-polynomial lower bound for learning nonparametric mixtures

Abstract

We study the problem of learning nonparametric distributions in a finite mixture, and establish a super-polynomial lower bound on the sample complexity of learning the component distributions in such models. Namely, we are given i.i.d. samples from ff where f=\sum_{i=1}^k w_i f_i, \quad\sum_{i=1}^k w_i=1, \quad w_i>0 and we are interested in learning each component fif_i. Without any assumptions on fif_i, this problem is ill-posed. In order to identify the components fif_i, we assume that each fif_i can be written as a convolution of a Gaussian and a compactly supported density νi\nu_i with supp(νi)supp(νj)=\text{supp}(\nu_i)\cap \text{supp}(\nu_j)=\emptyset. Our main result shows that Ω((1ε)Cloglog1ε)\Omega((\frac{1}{\varepsilon})^{C\log\log \frac{1}{\varepsilon}}) samples are required for estimating each fif_i. The proof relies on a fast rate for approximation with Gaussians, which may be of independent interest. This result has important implications for the hardness of learning more general nonparametric latent variable models that arise in machine learning applications.

View on arXiv
Comments on this paper