40
0
v1v2 (latest)

Choosing the pp in LpL_p loss: rate adaptivity on the symmetric location problem

Abstract

Given univariate random variables Y1,,YnY_1, \ldots, Y_n with the Uniform(θ01,θ0+1)\text{Uniform}(\theta_0 - 1, \theta_0 + 1) distribution, the sample midrange Y(n)+Y(1)2\frac{Y_{(n)}+Y_{(1)}}{2} is the MLE for θ0\theta_0 and estimates θ0\theta_0 with error of order 1/n1/n, which is much smaller compared with the 1/n1/\sqrt{n} error rate of the usual sample mean estimator. However, the sample midrange performs poorly when the data has say the Gaussian N(θ0,1)N(\theta_0, 1) distribution, with an error rate of 1/logn1/\sqrt{\log n}. In this paper, we propose an estimator of the location θ0\theta_0 with a rate of convergence that can, in many settings, adapt to the underlying distribution which we assume to be symmetric around θ0\theta_0 but is otherwise unknown. When the underlying distribution is compactly supported, we show that our estimator attains a rate of convergence of n1αn^{-\frac{1}{\alpha}} up to polylog factors, where the rate parameter α\alpha can take on any value in (0,2](0, 2] and depends on the moments of the underlying distribution. Our estimator is formed by the γ\ell^\gamma-center of the data, for a γ2\gamma\geq2 chosen in a data-driven way -- by minimizing a criterion motivated by the asymptotic variance. Our approach can be directly applied to the regression setting where θ0\theta_0 is a function of observed features and motivates the use of γ\ell^\gamma loss function for γ>2\gamma > 2 in certain settings.

View on arXiv
Comments on this paper