150

Maximum Likelihood Estimation for Learning Populations of Parameters

International Conference on Machine Learning (ICML), 2019
Abstract

Consider a setting with NN independent individuals, each with an unknown parameter, pi[0,1]p_i \in [0, 1] drawn from some unknown distribution PP^\star. After observing the outcomes of tt independent Bernoulli trials, i.e., XiBinomial(t,pi)X_i \sim \text{Binomial}(t, p_i) per individual, our objective is to accurately estimate PP^\star. This problem arises in numerous domains, including the social sciences, psychology, health-care, and biology, where the size of the population under study is usually large while the number of observations per individual is often limited. Our main result shows that, in the regime where tNt \ll N, the maximum likelihood estimator (MLE) is both statistically minimax optimal and efficiently computable. Precisely, for sufficiently large NN, the MLE achieves the information theoretic optimal error bound of O(1t)\mathcal{O}(\frac{1}{t}) for t<clogNt < c\log{N}, with regards to the earth mover's distance (between the estimated and true distributions). More generally, in an exponentially large interval of tt beyond clogNc \log{N}, the MLE achieves the minimax error bound of O(1tlogN)\mathcal{O}(\frac{1}{\sqrt{t\log N}}). In contrast, regardless of how large NN is, the naive "plug-in" estimator for this problem only achieves the sub-optimal error of Θ(1t)\Theta(\frac{1}{\sqrt{t}}).

View on arXiv
Comments on this paper