69

Computational-Statistical Tradeoffs from NP-hardness

Guy Blanc
Caleb Koch
Carmen Strassle
Li-Yang Tan
Main:26 Pages
5 Figures
Bibliography:4 Pages
1 Tables
Abstract

A central question in computer science and statistics is whether efficient algorithms can achieve the information-theoretic limits of statistical problems. Many computational-statistical tradeoffs have been shown under average-case assumptions, but since statistical problems are average-case in nature, it has been a challenge to base them on standard worst-case assumptions.In PAC learning where such tradeoffs were first studied, the question is whether computational efficiency can come at the cost of using more samples than information-theoretically necessary. We base such tradeoffs on NP\mathsf{NP}-hardness and obtain:\circ Sharp computational-statistical tradeoffs assuming NP\mathsf{NP} requires exponential time: For every polynomial p(n)p(n), there is an nn-variate class CC with VC dimension 11 such that the sample complexity of time-efficiently learning CC is Θ(p(n))\Theta(p(n)).\circ A characterization of RP\mathsf{RP} vs. NP\mathsf{NP} in terms of learning: RP=NP\mathsf{RP} = \mathsf{NP} iff every NP\mathsf{NP}-enumerable class is learnable with O(VCdim(C))O(\mathrm{VCdim}(C)) samples in polynomial time. The forward implication has been known since (Pitt and Valiant, 1988); we prove the reverse implication.Notably, all our lower bounds hold against improper learners. These are the first NP\mathsf{NP}-hardness results for improperly learning a subclass of polynomial-size circuits, circumventing formal barriers of Applebaum, Barak, and Xiao (2008).

View on arXiv
Comments on this paper