149

Regression for matrix-valued data via Kronecker products factorization

Abstract

We study the matrix-variate regression problem Yi=kβ1kXiβ2k+EiY_i = \sum_{k} \beta_{1k} X_i \beta_{2k}^{\top} + E_i for i=1,2,ni=1,2\dots,n in the high dimensional regime wherein the response YiY_i are matrices whose dimensions p1×p2p_{1}\times p_{2} outgrow both the sample size nn and the dimensions q1×q2q_{1}\times q_{2} of the predictor variables XiX_i i.e., q1,q2np1,p2q_{1},q_{2} \ll n \ll p_{1},p_{2}. We propose an estimation algorithm, termed KRO-PRO-FAC, for estimating the parameters {β1k}p1×q1\{\beta_{1k}\} \subset \Re^{p_1 \times q_1} and {β2k}p2×q2\{\beta_{2k}\} \subset \Re^{p_2 \times q_2} that utilizes the Kronecker product factorization and rearrangement operations from Van Loan and Pitsianis (1993). The KRO-PRO-FAC algorithm is computationally efficient as it does not require estimating the covariance between the entries of the {Yi}\{Y_i\}. We establish perturbation bounds between β^1kβ1k\hat{\beta}_{1k} -\beta_{1k} and β^2kβ2k\hat{\beta}_{2k} - \beta_{2k} in spectral norm for the setting where either the rows of EiE_i or the columns of EiE_i are independent sub-Gaussian random vectors. Numerical studies on simulated and real data indicate that our procedure is competitive, in terms of both estimation error and predictive accuracy, compared to other existing methods.

View on arXiv
Comments on this paper