15
1

Optimal Estimator for Linear Regression with Shuffled Labels

Abstract

This paper considers the task of linear regression with shuffled labels, i.e., Y=ΠXB+W\mathbf Y = \mathbf \Pi \mathbf X \mathbf B + \mathbf W, where YRn×m,PiRn×n,XRn×p,BRp×m\mathbf Y \in \mathbb R^{n\times m}, \mathbf Pi \in \mathbb R^{n\times n}, \mathbf X\in \mathbb R^{n\times p}, \mathbf B \in \mathbb R^{p\times m}, and WRn×m\mathbf W\in \mathbb R^{n\times m}, respectively, represent the sensing results, (unknown or missing) corresponding information, sensing matrix, signal of interest, and additive sensing noise. Given the observation Y\mathbf Y and sensing matrix X\mathbf X, we propose a one-step estimator to reconstruct (Π,B)(\mathbf \Pi, \mathbf B). From the computational perspective, our estimator's complexity is O(n3+np2m)O(n^3 + np^2m), which is no greater than the maximum complexity of a linear assignment algorithm (e.g., O(n3)O(n^3)) and a least square algorithm (e.g., O(np2m)O(np^2 m)). From the statistical perspective, we divide the minimum snrsnr requirement into four regimes, e.g., unknown, hard, medium, and easy regimes; and present sufficient conditions for the correct permutation recovery under each regime: (i)(i) snrΩ(1)snr \geq \Omega(1) in the easy regime; (ii)(ii) snrΩ(logn)snr \geq \Omega(\log n) in the medium regime; and (iii)(iii) snrΩ((logn)c0nc1/srank(B))snr \geq \Omega((\log n)^{c_0}\cdot n^{{c_1}/{srank(\mathbf B)}}) in the hard regime (c0,c1c_0, c_1 are some positive constants and srank(B)srank(\mathbf B) denotes the stable rank of B\mathbf B). In the end, we also provide numerical experiments to confirm the above claims.

View on arXiv
Comments on this paper