On the Power of Localized Perceptron for Label-Optimal Learning of Halfspaces with Adversarial Noise

We study {\em online} active learning of homogeneous -sparse halfspaces in with adversarial noise \cite{kearns1992toward}, where the overall probability of a noisy label is constrained to be at most and the marginal distribution over unlabeled data is unchanged. Our main contribution is a state-of-the-art online active learning algorithm that achieves near-optimal attribute efficiency, label and sample complexity under mild distributional assumptions. In particular, under the conditions that the marginal distribution is isotropic log-concave and , where is the target error rate, we show that our algorithm PAC learns the underlying halfspace in polynomial time with near-optimal label complexity bound of and sample complexity bound of . Prior to this work, existing online algorithms designed for tolerating the adversarial noise are either subject to label complexity polynomial in or , or work under the restrictive uniform marginal distribution. As an immediate corollary of our main result, we show that under the more challenging agnostic model \cite{kearns1992toward} where no assumption is made on the noise rate, our active learner achieves an error rate of with the same running time and label and sample complexity, where is the best possible error rate achievable by any homogeneous -sparse halfspace. Our algorithm builds upon the celebrated Perceptron while leveraging novel localized sampling and semi-random gradient update to tolerate the adversarial noise. We believe that our algorithmic design and analysis are of independent interest, and may shed light on learning halfspaces with broader noise models.
View on arXiv