369
v1v2v3v4 (latest)

Minimax risk of matrix denoising by singular value thresholding

Abstract

An unknown mm by nn matrix X0X_0 is to be estimated from noisy measurements Y=X0+ZY=X_0+Z, where the noise matrix ZZ has i.i.d. Gaussian entries. A popular matrix denoising scheme solves the nuclear norm penalization problem minXYXF2/2+λX\operatorname {min}_X\|Y-X\|_F^2/2+\lambda\|X\|_*, where X\|X\|_* denotes the nuclear norm (sum of singular values). This is the analog, for matrices, of 1\ell_1 penalization in the vector case. It has been empirically observed that if X0X_0 has low rank, it may be recovered quite accurately from the noisy measurement YY. In a proportional growth framework where the rank rnr_n, number of rows mnm_n and number of columns nn all tend to \infty proportionally to each other (rn/mnρr_n/m_n\rightarrow \rho, mn/nβm_n/n\rightarrow \beta), we evaluate the asymptotic minimax MSE M(ρ,β)=limmn,ninfλsuprank(X)rnMSE(X0,X^λ)\mathcal {M}(\rho,\beta)=\lim_{m_n,n\rightarrow \infty}\inf_{\lambda}\sup_{\operatorname {rank}(X)\leq r_n}\operatorname {MSE}(X_0,\hat{X}_{\lambda}). Our formulas involve incomplete moments of the quarter- and semi-circle laws (β=1\beta=1, square case) and the Mar\v{c}enko-Pastur law (β<1\beta<1, nonsquare case). For finite mm and nn, we show that MSE increases as the nonzero singular values of X0X_0 grow larger. As a result, the finite-nn worst-case MSE, a quantity which can be evaluated numerically, is achieved when the signal X0X_0 is "infinitely strong." The nuclear norm penalization problem is solved by applying soft thresholding to the singular values of YY. We also derive the minimax threshold, namely the value λ(ρ)\lambda^*(\rho), which is the optimal place to threshold the singular values. All these results are obtained for general (nonsquare, nonsymmetric) real matrices. Comparable results are obtained for square symmetric nonnegative-definite matrices.

View on arXiv
Comments on this paper