143
v1v2 (latest)

High Dimensional Linear Regression using Lattice Basis Reduction

Abstract

We consider a high dimensional linear regression problem where the goal is to efficiently recover an unknown vector β\beta^* from nn noisy linear observations Y=Xβ+WRnY=X\beta^*+W \in \mathbb{R}^n, for known XRn×pX \in \mathbb{R}^{n \times p} and unknown WRnW \in \mathbb{R}^n. Unlike most of the literature on this model we make no sparsity assumption on β\beta^*. Instead we adopt a regularization based on assuming that the underlying vectors β\beta^* have rational entries with the same denominator QZ>0Q \in \mathbb{Z}_{>0}. We call this QQ-rationality assumption. We propose a new polynomial-time algorithm for this task which is based on the seminal Lenstra-Lenstra-Lovasz (LLL) lattice basis reduction algorithm. We establish that under the QQ-rationality assumption, our algorithm recovers exactly the vector β\beta^* for a large class of distributions for the iid entries of XX and non-zero noise WW. We prove that it is successful under small noise, even when the learner has access to only one observation (n=1n=1). Furthermore, we prove that in the case of the Gaussian white noise for WW, n=o(p/logp)n=o\left(p/\log p\right) and QQ sufficiently large, our algorithm tolerates a nearly optimal information-theoretic level of the noise.

View on arXiv
Comments on this paper