66
24
v1v2 (latest)

Near-Optimal Sample Complexity Bounds for Maximum Likelihood Estimation of Multivariate Log-concave Densities

Abstract

We study the problem of learning multivariate log-concave densities with respect to a global loss function. We obtain the first upper bound on the sample complexity of the maximum likelihood estimator (MLE) for a log-concave density on Rd\mathbb{R}^d, for all d4d \geq 4. Prior to this work, no finite sample upper bound was known for this estimator in more than 33 dimensions. In more detail, we prove that for any d1d \geq 1 and ϵ>0\epsilon>0, given O~d((1/ϵ)(d+3)/2)\tilde{O}_d((1/\epsilon)^{(d+3)/2}) samples drawn from an unknown log-concave density f0f_0 on Rd\mathbb{R}^d, the MLE outputs a hypothesis hh that with high probability is ϵ\epsilon-close to f0f_0, in squared Hellinger loss. A sample complexity lower bound of Ωd((1/ϵ)(d+1)/2)\Omega_d((1/\epsilon)^{(d+1)/2}) was previously known for any learning algorithm that achieves this guarantee. We thus establish that the sample complexity of the log-concave MLE is near-optimal, up to an O~(1/ϵ)\tilde{O}(1/\epsilon) factor.

View on arXiv
Comments on this paper