106
65

Simultaneous support recovery in high dimensions: Benefits and perils of block 1/\ell_1/\ell_\infty-regularization

Abstract

Consider the use of 1/\ell_{1}/\ell_{\infty}-regularized regression for joint estimation of a \pdim×\numreg\pdim \times \numreg matrix of regression coefficients. We analyze the high-dimensional scaling of 1/\ell_1/\ell_\infty-regularized quadratic programming, considering both consistency in \ell_\infty-norm, and variable selection. We begin by establishing bounds on the \ell_\infty-error as well sufficient conditions for exact variable selection for fixed and random designs. Our second set of results applies to \numreg=2\numreg = 2 linear regression problems with standard Gaussian designs whose supports overlap in a fraction α[0,1]\alpha \in [0,1] of their entries: for this problem class, we prove that the 1/\ell_{1}/\ell_{\infty}-regularized method undergoes a phase transition--that is, a sharp change from failure to success--characterized by the rescaled sample size θ1,(n,p,s,α)=n/{(43α)slog(p(2α)s)}\theta_{1,\infty}(n, p, s, \alpha) = n/\{(4 - 3 \alpha) s \log(p-(2- \alpha) s)\}. An implication of this threshold is that use of 1/\ell_1 / \ell_{\infty}-regularization yields improved statistical efficiency if the overlap parameter is large enough (α>2/3\alpha > 2/3), but has \emph{worse} statistical efficiency than a naive Lasso-based approach for moderate to small overlap (α<2/3\alpha < 2/3). These results indicate that some caution needs to be exercised in the application of 1/\ell_1/\ell_\infty block regularization: if the data does not match its structure closely enough, it can impair statistical performance relative to computationally less expensive schemes.

View on arXiv
Comments on this paper