10
3

Chi-square and normal inference in high-dimensional multi-task regression

Abstract

The paper proposes chi-square and normal inference methodologies for the unknown coefficient matrix BB^* of size p×Tp\times T in a Multi-Task (MT) linear model with pp covariates, TT tasks and nn observations under a row-sparse assumption on BB^*. The row-sparsity ss, dimension pp and number of tasks TT are allowed to grow with nn. In the high-dimensional regime pnp\ggg n, in order to leverage row-sparsity, the MT Lasso is considered. We build upon the MT Lasso with a de-biasing scheme to correct for the bias induced by the penalty. This scheme requires the introduction of a new data-driven object, coined the interaction matrix, that captures effective correlations between noise vector and residuals on different tasks. This matrix is psd, of size T×TT\times T and can be computed efficiently. The interaction matrix lets us derive asymptotic normal and χT2\chi^2_T results under Gaussian design and sT+slog(p/s)n0\frac{sT+s\log(p/s)}{n}\to0 which corresponds to consistency in Frobenius norm. These asymptotic distribution results yield valid confidence intervals for single entries of BB^* and valid confidence ellipsoids for single rows of BB^*, for both known and unknown design covariance Σ\Sigma. While previous proposals in grouped-variables regression require row-sparsity sns\lesssim\sqrt n up to constants depending on TT and logarithmic factors in n,pn,p, the de-biasing scheme using the interaction matrix provides confidence intervals and χT2\chi^2_T confidence ellipsoids under the conditions min(T2,log8p)/n0{\min(T^2,\log^8p)}/{n}\to 0 and \frac{sT+s\log(p/s)+\|\Sigma^{-1}e_j\|_0\log p}{n}\to0, \quad \frac{\min(s,\|\Sigma^{-1}e_j\|_0)}{\sqrt n} \sqrt{[T+\log(p/s)]\log p}\to 0, allowing row-sparsity sns\ggg\sqrt n when Σ1ej0Tn\|\Sigma^{-1}e_j\|_0 \sqrt T\lll \sqrt{n} up to logarithmic factors.

View on arXiv
Comments on this paper