14
6

The Complexity of Sparse Tensor PCA

Abstract

We study the problem of sparse tensor principal component analysis: given a tensor Y=W+λxp\pmb Y = \pmb W + \lambda x^{\otimes p} with WpRn\pmb W \in \otimes^p\mathbb{R}^n having i.i.d. Gaussian entries, the goal is to recover the kk-sparse unit vector xRnx \in \mathbb{R}^n. The model captures both sparse PCA (in its Wigner form) and tensor PCA. For the highly sparse regime of knk \leq \sqrt{n}, we present a family of algorithms that smoothly interpolates between a simple polynomial-time algorithm and the exponential-time exhaustive search algorithm. For any 1tk1 \leq t \leq k, our algorithms recovers the sparse vector for signal-to-noise ratio λO~(t(k/t)p/2)\lambda \geq \tilde{\mathcal{O}} (\sqrt{t} \cdot (k/t)^{p/2}) in time O~(np+t)\tilde{\mathcal{O}}(n^{p+t}), capturing the state-of-the-art guarantees for the matrix settings (in both the polynomial-time and sub-exponential time regimes). Our results naturally extend to the case of rr distinct kk-sparse signals with disjoint supports, with guarantees that are independent of the number of spikes. Even in the restricted case of sparse PCA, known algorithms only recover the sparse vectors for λO~(kr)\lambda \geq \tilde{\mathcal{O}}(k \cdot r) while our algorithms require λO~(k)\lambda \geq \tilde{\mathcal{O}}(k). Finally, by analyzing the low-degree likelihood ratio, we complement these algorithmic results with rigorous evidence illustrating the trade-offs between signal-to-noise ratio and running time. This lower bound captures the known lower bounds for both sparse PCA and tensor PCA. In this general model, we observe a more intricate three-way trade-off between the number of samples nn, the sparsity kk, and the tensor power pp.

View on arXiv
Comments on this paper