ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2007.08137
20
35

Optimal Robust Linear Regression in Nearly Linear Time

16 July 2020
Yeshwanth Cherapanamjeri
Efe Aras
Nilesh Tripuraneni
Michael I. Jordan
Nicolas Flammarion
Peter L. Bartlett
ArXivPDFHTML
Abstract

We study the problem of high-dimensional robust linear regression where a learner is given access to nnn samples from the generative model Y=⟨X,w∗⟩+ϵY = \langle X,w^* \rangle + \epsilonY=⟨X,w∗⟩+ϵ (with X∈RdX \in \mathbb{R}^dX∈Rd and ϵ\epsilonϵ independent), in which an η\etaη fraction of the samples have been adversarially corrupted. We propose estimators for this problem under two settings: (i) XXX is L4-L2 hypercontractive, E[XX⊤]\mathbb{E} [XX^\top]E[XX⊤] has bounded condition number and ϵ\epsilonϵ has bounded variance and (ii) XXX is sub-Gaussian with identity second moment and ϵ\epsilonϵ is sub-Gaussian. In both settings, our estimators: (a) Achieve optimal sample complexities and recovery guarantees up to log factors and (b) Run in near linear time (O~(nd/η6)\tilde{O}(nd / \eta^6)O~(nd/η6)). Prior to our work, polynomial time algorithms achieving near optimal sample complexities were only known in the setting where XXX is Gaussian with identity covariance and ϵ\epsilonϵ is Gaussian, and no linear time estimators were known for robust linear regression in any setting. Our estimators and their analysis leverage recent developments in the construction of faster algorithms for robust mean estimation to improve runtimes, and refined concentration of measure arguments alongside Gaussian rounding techniques to improve statistical sample complexities.

View on arXiv
Comments on this paper