Robust learning of halfspaces under log-concave marginals
We say that a classifier is \emph{adversarially robust} to perturbations of norm if, with high probability over a point drawn from the input distribution, there is no point within distance from that is classified differently. The \emph{boundary volume} is the probability that a point falls within distance of a point with a different label. This work studies the task of computationally efficient learning of hypotheses with small boundary volume, where the input is distributed as a subgaussian isotropic log-concave distribution over .Linear threshold functions are adversarially robust; they have boundary volume proportional to . Such concept classes are efficiently learnable by polynomial regression, which produces a polynomial threshold function (PTF), but PTFs in general may have boundary volume , even for .We give an algorithm that agnostically learns linear threshold functions and returns a classifier with boundary volume at radius of perturbation . The time and sample complexity of matches the complexity of polynomial regression.Our algorithm augments the classic approach of polynomial regression with three additional steps: a) performing the -error regression under noise sensitivity constraints, b) a structured partitioning and rounding step that returns a Boolean classifier with error and noise sensitivity simultaneously, and c) a local corrector that ``smooths'' a function with low noise sensitivity into a function that is adversarially robust.
View on arXiv