ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2206.08092
17
5

On the well-spread property and its relation to linear regression

16 June 2022
Hongjie Chen
Tommaso dÓrsi
ArXivPDFHTML
Abstract

We consider the robust linear regression model y=Xβ∗+η\boldsymbol{y} = X\beta^* + \boldsymbol{\eta}y=Xβ∗+η, where an adversary oblivious to the design X∈Rn×dX \in \mathbb{R}^{n \times d}X∈Rn×d may choose η\boldsymbol{\eta}η to corrupt all but a (possibly vanishing) fraction of the observations y\boldsymbol{y}y in an arbitrary way. Recent work [dLN+21, dNS21] has introduced efficient algorithms for consistent recovery of the parameter vector. These algorithms crucially rely on the design matrix being well-spread (a matrix is well-spread if its column span is far from any sparse vector). In this paper, we show that there exists a family of design matrices lacking well-spreadness such that consistent recovery of the parameter vector in the above robust linear regression model is information-theoretically impossible. We further investigate the average-case time complexity of certifying well-spreadness of random matrices. We show that it is possible to efficiently certify whether a given nnn-by-ddd Gaussian matrix is well-spread if the number of observations is quadratic in the ambient dimension. We complement this result by showing rigorous evidence -- in the form of a lower bound against low-degree polynomials -- of the computational hardness of this same certification problem when the number of observations is o(d2)o(d^2)o(d2).

View on arXiv
Comments on this paper