ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2002.04840
95
44
v1v2v3 (latest)

Efficient active learning of sparse halfspaces with arbitrary bounded noise

12 February 2020
Chicheng Zhang
Jie Shen
Pranjal Awasthi
ArXiv (abs)PDFHTML
Abstract

In this work we study active learning of homogeneous sss-sparse halfspaces in Rd\mathbb{R}^dRd under label noise. Even in the absence of label noise this is a challenging problem and only recently have label complexity bounds of the form O~(s⋅polylog(d,1ϵ))\tilde{O} \left(s \cdot \mathrm{polylog}(d, \frac{1}{\epsilon}) \right)O~(s⋅polylog(d,ϵ1​)) been established in \citet{zhang2018efficient} for computationally efficient algorithms under the broad class of isotropic log-concave distributions. In contrast, under high levels of label noise, the label complexity bounds achieved by computationally efficient algorithms are much worse. When the label noise satisfies the {\em Massart} condition~\citep{massart2006risk}, i.e., each label is flipped with probability at most η\etaη for a parameter η∈[0,12)\eta \in [0,\frac 1 2)η∈[0,21​), the work of \citet{awasthi2016learning} provides a computationally efficient active learning algorithm under isotropic log-concave distributions with label complexity O~(spoly(1/(1−2η))poly(log⁡d,1ϵ))\tilde{O} \left(s^{\mathrm{poly}{(1/(1-2\eta))}} \mathrm{poly}(\log d, \frac{1}{\epsilon}) \right)O~(spoly(1/(1−2η))poly(logd,ϵ1​)). Hence the algorithm is label-efficient only when the noise rate η\etaη is a constant. In this work, we substantially improve on the state of the art by designing a polynomial time algorithm for active learning of sss-sparse halfspaces under bounded noise and isotropic log-concave distributions, with a label complexity of O~(s(1−2η)4polylog(d,1ϵ))\tilde{O} \left(\frac{s}{(1-2\eta)^4} \mathrm{polylog} (d, \frac 1 \epsilon) \right)O~((1−2η)4s​polylog(d,ϵ1​)). Hence, our new algorithm is label-efficient even for noise rates close to 12\frac{1}{2}21​. Prior to our work, such a result was not known even for the random classification noise model. Our algorithm builds upon existing margin-based algorithmic framework and at each iteration performs a sequence of online mirror descent updates on a carefully chosen loss sequence, and uses a novel gradient update rule that accounts for the bounded noise.

View on arXiv
Comments on this paper