Beyond Catoni: Sharper Rates for Heavy-Tailed and Robust Mean Estimation

We study the fundamental problem of estimating the mean of a -dimensional distribution with covariance given samples. When , \cite{catoni} showed an estimator with error , with probability , matching the Gaussian error rate. For , a natural estimator outputs the center of the minimum enclosing ball of one-dimensional confidence intervals to achieve a confidence radius of , incurring a -factor loss over the Gaussian rate. When the term dominates by a factor, \cite{lee2022optimal-highdim} showed an improved estimator matching the Gaussian rate. This raises a natural question: Is the loss \emph{necessary} when the term dominates? We show that the answer is \emph{no} -- we construct an estimator that improves over the above naive estimator by a constant factor. We also consider robust estimation, where an adversary is allowed to corrupt an -fraction of samples arbitrarily: in this case, we show that the above strategy of combining one-dimensional estimates and incurring the -factor \emph{is} optimal in the infinite-sample limit.
View on arXiv