Fast learning rates with heavy-tailed losses

We study fast learning rates when the losses are not necessarily bounded and may have a distribution with heavy tails. To enable such analyses, we introduce two new conditions: (i) the envelope function , where is the loss function and is the hypothesis class, exists and is -integrable, and (ii) satisfies the multi-scale Bernstein's condition on . Under these assumptions, we prove that learning rate faster than can be obtained and, depending on and the multi-scale Bernstein's powers, can be arbitrarily close to . We then verify these assumptions and derive fast learning rates for the problem of vector quantization by -means clustering with heavy-tailed distributions. The analyses enable us to obtain novel learning rates that extend and complement existing results in the literature from both theoretical and practical viewpoints.
View on arXiv