Chi-square and normal inference in high-dimensional multi-task regression

The paper proposes chi-square and normal inference methodologies for the unknown coefficient matrix of size in a Multi-Task (MT) linear model with covariates, tasks and observations under a row-sparse assumption on . The row-sparsity , dimension and number of tasks are allowed to grow with . In the high-dimensional regime , in order to leverage row-sparsity, the MT Lasso is considered. We build upon the MT Lasso with a de-biasing scheme to correct for the bias induced by the penalty. This scheme requires the introduction of a new data-driven object, coined the interaction matrix, that captures effective correlations between noise vector and residuals on different tasks. This matrix is psd, of size and can be computed efficiently. The interaction matrix lets us derive asymptotic normal and results under Gaussian design and which corresponds to consistency in Frobenius norm. These asymptotic distribution results yield valid confidence intervals for single entries of and valid confidence ellipsoids for single rows of , for both known and unknown design covariance . While previous proposals in grouped-variables regression require row-sparsity up to constants depending on and logarithmic factors in , the de-biasing scheme using the interaction matrix provides confidence intervals and confidence ellipsoids under the conditions and \frac{sT+s\log(p/s)+\|\Sigma^{-1}e_j\|_0\log p}{n}\to0, \quad \frac{\min(s,\|\Sigma^{-1}e_j\|_0)}{\sqrt n} \sqrt{[T+\log(p/s)]\log p}\to 0, allowing row-sparsity when up to logarithmic factors.
View on arXiv