257

Reliable Learning of Halfspaces under Gaussian Marginals

Neural Information Processing Systems (NeurIPS), 2024
Abstract

We study the problem of PAC learning halfspaces in the reliable agnostic model of Kalai et al. (2012). The reliable PAC model captures learning scenarios where one type of error is costlier than the others. Our main positive result is a new algorithm for reliable learning of Gaussian halfspaces on Rd\mathbb{R}^d with sample and computational complexity dO(log(min{1/α,1/ϵ}))min(2log(1/ϵ)O(log(1/α)),2poly(1/ϵ))  ,d^{O(\log (\min\{1/\alpha, 1/\epsilon\}))}\min (2^{\log(1/\epsilon)^{O(\log (1/\alpha))}},2^{\mathrm{poly}(1/\epsilon)})\;, where ϵ\epsilon is the excess error and α\alpha is the bias of the optimal halfspace. We complement our upper bound with a Statistical Query lower bound suggesting that the dΩ(log(1/α))d^{\Omega(\log (1/\alpha))} dependence is best possible. Conceptually, our results imply a strong computational separation between reliable agnostic learning and standard agnostic learning of halfspaces in the Gaussian setting.

View on arXiv
Main:12 Pages
Bibliography:4 Pages
Appendix:19 Pages
Comments on this paper