40

Robust learning of halfspaces under log-concave marginals

Main:9 Pages
Bibliography:3 Pages
Appendix:26 Pages
Abstract

We say that a classifier is \emph{adversarially robust} to perturbations of norm rr if, with high probability over a point xx drawn from the input distribution, there is no point within distance r\le r from xx that is classified differently. The \emph{boundary volume} is the probability that a point falls within distance rr of a point with a different label. This work studies the task of computationally efficient learning of hypotheses with small boundary volume, where the input is distributed as a subgaussian isotropic log-concave distribution over Rd\mathbb{R}^d.Linear threshold functions are adversarially robust; they have boundary volume proportional to rr. Such concept classes are efficiently learnable by polynomial regression, which produces a polynomial threshold function (PTF), but PTFs in general may have boundary volume Ω(1)\Omega(1), even for r1r \ll 1.We give an algorithm that agnostically learns linear threshold functions and returns a classifier with boundary volume O(r+ε)O(r+\varepsilon) at radius of perturbation rr. The time and sample complexity of dO~(1/ε2)d^{\tilde{O}(1/\varepsilon^2)} matches the complexity of polynomial regression.Our algorithm augments the classic approach of polynomial regression with three additional steps: a) performing the 1\ell_1-error regression under noise sensitivity constraints, b) a structured partitioning and rounding step that returns a Boolean classifier with error opt+O(ε)\textsf{opt} + O(\varepsilon) and noise sensitivity O(r+ε)O(r+\varepsilon) simultaneously, and c) a local corrector that ``smooths'' a function with low noise sensitivity into a function that is adversarially robust.

View on arXiv
Comments on this paper