Optimal Spectral Recovery of a Planted Vector in a Subspace

Recovering a planted vector in an -dimensional random subspace of is a generic task related to many problems in machine learning and statistics, such as dictionary learning, subspace recovery, principal component analysis, and non-Gaussian component analysis. In this work, we study computationally efficient estimation and detection of a planted vector whose norm differs from that of a Gaussian vector with the same norm. For instance, in the special case where is an -sparse vector with Bernoulli-Gaussian or Bernoulli-Rademacher entries, our results include the following: (1) We give an improved analysis of a slight variant of the spectral method proposed by Hopkins, Schramm, Shi, and Steurer (2016), showing that it approximately recovers with high probability in the regime . This condition subsumes the conditions or required by previous work up to polylogarithmic factors. We achieve error bounds for the spectral estimator via a leave-one-out analysis, from which it follows that a simple thresholding procedure exactly recovers with Bernoulli-Rademacher entries, even in the dense case . (2) We study the associated detection problem and show that in the regime , any spectral method from a large class (and more generally, any low-degree polynomial of the input) fails to detect the planted vector. This matches the condition for recovery and offers evidence that no polynomial-time algorithm can succeed in recovering a Bernoulli-Gaussian vector when .
View on arXiv