On Semi-supervised Estimation of Discrete Distributions under
f-divergences
We study the problem of estimating the joint probability mass function (pmf) over two random variables. In particular, the estimation is based on the observation of samples containing both variables and samples missing one fixed variable. We adopt the minimax framework with loss functions. Recent work established that univariate minimax estimator combinations achieve minimax risk with the optimal first-order constant for in the regime , questions remained for and various -divergences. In our study, we affirm that these composite estimators are indeed minimax optimal for loss functions, specifically for the range , including the critical loss. Additionally, we ascertain their optimality for a suite of -divergences, such as KL, , Squared Hellinger, and Le Cam divergences.
View on arXiv