193

Sharper lower bounds on the performance of the empirical risk minimization algorithm

Abstract

We present an argument based on the multidimensional and the uniform central limit theorems, proving that, under some geometrical assumptions between the target function TT and the learning class FF, the excess risk of the empirical risk minimization algorithm is lower bounded by \[\frac{\mathbb{E}\sup_{q\in Q}G_q}{\sqrt{n}}\delta,\] where (Gq)qQ(G_q)_{q\in Q} is a canonical Gaussian process associated with QQ (a well chosen subset of FF) and δ\delta is a parameter governing the oscillations of the empirical excess risk function over a small ball in FF.

View on arXiv
Comments on this paper