67

Noise-Augmented 0\ell_0 Regularization of Tensor Regression with Tucker Decomposition

Abstract

Tensor data are multi-dimension arrays. Low-rank decomposition-based regression methods with tensor predictors exploit the structural information in tensor predictors while significantly reducing the number of parameters in tensor regression. We propose a method named NA0_0CT2^2 (Noise Augmentation for 0\ell_0 regularization on Core Tensor in Tucker decomposition) to regularize the parameters in tensor regression (TR), coupled with Tucker decomposition. We establish theoretically that NA0_0CT2^2 achieves exact 0\ell_0 regularization in linear TR and generalized linear TR on the core tensor from the Tucker decomposition. To our knowledge, NA0_0CT2^2 is the first Tucker decomposition-based regularization method in TR to achieve 0\ell_0 in core tensor. NA0_0CT2^2 is implemented through an iterative procedure and involves two simple steps in each iteration -- generating noisy data based on the core tensor from the Tucker decomposition of the updated parameter estimate and running a regular GLM on noise-augmented data on vectorized predictors. We demonstrate the implementation of NA0_0CT2^2 and its 0\ell_0 regularization effect in both simulation studies and real data applications. The results suggest that NA0_0CT2^2 improves predictions compared to other decomposition-based TR approaches, with or without regularization and it also helps to identify important predictors though not designed for that purpose.

View on arXiv
Comments on this paper