61
25

Optimality of Maximum Likelihood for Log-Concave Density Estimation and Bounded Convex Regression

Abstract

In this paper, we study two fundamentals problems: estimation of a dd-dimensional log-concave distribution and bounded multivariate convex regression with random design. First, we show that for all d4d \ge 4 the maximum likelihood estimators of both problems achieve an optimal risk (up to a logarithmic factor) of Θd(n2/(d+1))\Theta_d(n^{-2/(d+1)}) in terms of squared Hellinger distance and L2L_2 squared distance, respectively. Previously, the optimality of both these estimators was known only for d3d\le 3. We also prove that the ϵ\epsilon-entropy numbers of the two aforementioned families are equal up to logarithmic factors. We complement these results by proving a sharp bound Θd(n2/(d+4))\Theta_d(n^{-2/(d+4)}) on the minimax rate (up to logarithmic factors) with respect to the total variation distance. Finally, we prove that estimating a log-concave density---even a uniform distribution on a convex set---up to a fixed accuracy requires \emph{at least} a number of samples which is exponential in the dimension. We do that by improving the dimensional constant in the best known lower bound for the minimax rate from 2dn2/(d+1)2^{-d}\cdot n^{-2/(d+1)} to cn2/(d+1)c\cdot n^{-2/(d+1)}.

View on arXiv
Comments on this paper