331

Effective Dimension of Exp-concave Optimization

Abstract

We investigate the role of the effective (a.k.a. statistical) dimension in determining both the statistical and the computational costs associated with exp-concave stochastic minimization. We derive sample complexity bounds that scale with dλϵ\frac{d_{\lambda}}{\epsilon}, where dλd_{\lambda} is the effective dimension associated with the regularization parameter λ\lambda. These are the first fast rates in this setting that do not exhibit any explicit dependence either on the intrinsic dimension or the 2\ell_{2}-norm of the optimal classifier. We also propose fast preconditioned methods that solve the ERM problem in time O~(nnz(X)+minλλλλ dλ2d)\tilde{O} \left(nnz(X)+\min_{\lambda'\ge\lambda}\frac{\lambda'}{\lambda}~d_{\lambda'}^{2}d \right), where nnz(X)nnz(X) is the number of nonzero entries in the data. Our analysis emphasizes interesting connections between leverage scores, algorithmic stability and regularization. In particular, our algorithm involves a novel technique for choosing a regularization parameter λ\lambda' that minimizes the complexity bound λλdλ2d\frac{\lambda'}{\lambda}\,d_{\lambda'}^{2}d, while avoiding the entire (approximate) computation of the effective dimension for each candidate λ\lambda'. All of our result extend to the kernel setting.

View on arXiv
Comments on this paper