192

Nearly Minimax Discrete Distribution Estimation in Kullback-Leibler Divergence with High Probability

Main:13 Pages
Bibliography:4 Pages
Appendix:21 Pages
Abstract

We consider the problem of estimating a discrete distribution pp with support of size KK and provide both upper and lower bounds with high probability in KL divergence. We prove that in the worst case, for any estimator p^\widehat{p}, with probability at least δ\delta, $\text{KL}(p \| \widehat{p}) \geq C\max\{K,\ln(K)\ln(1/\delta) \}/n $, where nn is the sample size and C>0C > 0 is a constant. We introduce a computationally efficient estimator pOTBp^{\text{OTB}}, based on Online to Batch conversion and suffix averaging, and show that with probability at least 1δ1 - \delta KL(pp^)C(Klog(log(K))+ln(K)ln(1/δ))/n\text{KL}(p \| \widehat{p}) \leq C(K\log(\log(K)) + \ln(K)\ln(1/\delta)) /n.

View on arXiv
Comments on this paper