487

Minimax rates of estimation for high-dimensional linear regression over q\ell_q-balls

Abstract

Consider the standard linear regression model \y=\Xmat\betastar+w\y = \Xmat \betastar + w, where \y\numobs\y \in \real^\numobs is an observation vector, \Xmat\numobs×\pdim\Xmat \in \real^{\numobs \times \pdim} is a design matrix, \betastar\pdim\betastar \in \real^\pdim is the unknown regression vector, and wN(0,σ2I)w \sim \mathcal{N}(0, \sigma^2 I) is additive Gaussian noise. This paper studies the minimax rates of convergence for estimation of \betastar\betastar for \rpar\ell_\rpar-losses and in the 2\ell_2-prediction loss, assuming that \betastar\betastar belongs to an \qpar\ell_{\qpar}-ball \Ballq(\myrad)\Ballq(\myrad) for some \qpar[0,1]\qpar \in [0,1]. We show that under suitable regularity conditions on the design matrix \Xmat\Xmat, the minimax error in 2\ell_2-loss and 2\ell_2-prediction loss scales as \Rq(log\pdimn)1\qpar2\Rq \big(\frac{\log \pdim}{n}\big)^{1-\frac{\qpar}{2}}. In addition, we provide lower bounds on minimax risks in \rpar\ell_{\rpar}-norms, for all \rpar[1,+],\rpar\qpar\rpar \in [1, +\infty], \rpar \neq \qpar. Our proofs of the lower bounds are information-theoretic in nature, based on Fano's inequality and results on the metric entropy of the balls \Ballq(\myrad)\Ballq(\myrad), whereas our proofs of the upper bounds are direct and constructive, involving direct analysis of least-squares over \qpar\ell_{\qpar}-balls. For the special case q=0q = 0, a comparison with 2\ell_2-risks achieved by computationally efficient 1\ell_1-relaxations reveals that although such methods can achieve the minimax rates up to constant factors, they require slightly stronger assumptions on the design matrix \Xmat\Xmat than algorithms involving least-squares over the 0\ell_0-ball.

View on arXiv
Comments on this paper