11
15

On the Power of Truncated SVD for General High-rank Matrix Estimation Problems

S. Du
Yining Wang
Aarti Singh
Abstract

We show that given an estimate A^\widehat{A} that is close to a general high-rank positive semi-definite (PSD) matrix AA in spectral norm (i.e., A^A2δ\|\widehat{A}-A\|_2 \leq \delta), the simple truncated SVD of A^\widehat{A} produces a multiplicative approximation of AA in Frobenius norm. This observation leads to many interesting results on general high-rank matrix estimation problems, which we briefly summarize below (AA is an n×nn\times n high-rank PSD matrix and AkA_k is the best rank-kk approximation of AA): (1) High-rank matrix completion: By observing Ω(nmax{ϵ4,k2}μ02AF2lognσk+1(A)2)\Omega(\frac{n\max\{\epsilon^{-4},k^2\}\mu_0^2\|A\|_F^2\log n}{\sigma_{k+1}(A)^2}) elements of AA where σk+1(A)\sigma_{k+1}\left(A\right) is the (k+1)\left(k+1\right)-th singular value of AA and μ0\mu_0 is the incoherence, the truncated SVD on a zero-filled matrix satisfies A^kAF(1+O(ϵ))AAkF\|\widehat{A}_k-A\|_F \leq (1+O(\epsilon))\|A-A_k\|_F with high probability. (2)High-rank matrix de-noising: Let A^=A+E\widehat{A}=A+E where EE is a Gaussian random noise matrix with zero mean and ν2/n\nu^2/n variance on each entry. Then the truncated SVD of A^\widehat{A} satisfies A^kAF(1+O(ν/σk+1(A)))AAkF+O(kν)\|\widehat{A}_k-A\|_F \leq (1+O(\sqrt{\nu/\sigma_{k+1}(A)}))\|A-A_k\|_F + O(\sqrt{k}\nu). (3) Low-rank Estimation of high-dimensional covariance: Given NN i.i.d.~samples X1,,XNNn(0,A)X_1,\cdots,X_N\sim\mathcal N_n(0,A), can we estimate AA with a relative-error Frobenius norm bound? We show that if N=Ω(nmax{ϵ4,k2}γk(A)2logN)N = \Omega\left(n\max\{\epsilon^{-4},k^2\}\gamma_k(A)^2\log N\right) for γk(A)=σ1(A)/σk+1(A)\gamma_k(A)=\sigma_1(A)/\sigma_{k+1}(A), then A^kAF(1+O(ϵ))AAkF\|\widehat{A}_k-A\|_F \leq (1+O(\epsilon))\|A-A_k\|_F with high probability, where A^=1Ni=1NXiXi\widehat{A}=\frac{1}{N}\sum_{i=1}^N{X_iX_i^\top} is the sample covariance.

View on arXiv
Comments on this paper