Consider a setting with independent individuals, each with an unknown parameter, drawn from some unknown distribution . After observing the outcomes of independent Bernoulli trials, i.e., per individual, our objective is to accurately estimate . This problem arises in numerous domains, including the social sciences, psychology, health-care, and biology, where the size of the population under study is usually large while the number of observations per individual is often limited. Our main result shows that, in the regime where , the maximum likelihood estimator (MLE) is both statistically minimax optimal and efficiently computable. Precisely, for sufficiently large , the MLE achieves the information theoretic optimal error bound of for , with regards to the earth mover's distance (between the estimated and true distributions). More generally, in an exponentially large interval of beyond , the MLE achieves the minimax error bound of . In contrast, regardless of how large is, the naive "plug-in" estimator for this problem only achieves the sub-optimal error of .
View on arXiv