56
1

Testing and reconstruction via decision trees

Abstract

We study sublinear and local computation algorithms for decision trees, focusing on testing and reconstruction. Our first result is a tester that runs in poly(logs,1/ε)nlogn\mathrm{poly}(\log s, 1/\varepsilon)\cdot n\log n time, makes poly(logs,1/ε)logn\mathrm{poly}(\log s,1/\varepsilon)\cdot \log n queries to an unknown function ff, and: \circ Accepts if ff is ε\varepsilon-close to a size-ss decision tree; \circ Rejects if ff is Ω(ε)\Omega(\varepsilon)-far from decision trees of size sO~((logs)2/ε2)s^{\tilde{O}((\log s)^2/\varepsilon^2)}. Existing testers distinguish size-ss decision trees from those that are ε\varepsilon-far from from size-ss decision trees in poly(ss,1/ε)n\mathrm{poly}(s^s,1/\varepsilon)\cdot n time with O~(s/ε)\tilde{O}(s/\varepsilon) queries. We therefore solve an incomparable problem, but achieve doubly-exponential-in-ss and exponential-in-ss improvements in time and query complexities respectively. We obtain our tester by designing a reconstruction algorithm for decision trees: given query access to a function ff that is close to a small decision tree, this algorithm provides fast query access to a small decision tree that is close to ff. By known relationships, our results yield reconstruction algorithms for numerous other boolean function properties -- Fourier degree, randomized and quantum query complexities, certificate complexity, sensitivity, etc. -- which in turn yield new testers for these properties. Finally, we give a hardness result for testing whether an unknown function is ε\varepsilon-close-to or Ω(ε)\Omega(\varepsilon)-far-from size-ss decision trees. We show that an efficient algorithm for this task would yield an efficient algorithm for properly learning decision trees, a central open problem of learning theory. It has long been known that proper learning algorithms for any class H\mathcal{H} yield property testers for H\mathcal{H}; this provides an example of a converse.

View on arXiv
Comments on this paper