Kernel Thinning
We introduce kernel thinning, a simple algorithm for generating better-than-Monte-Carlo approximations to distributions on . Given input points, a suitable reproducing kernel , and time, kernel thinning returns points with comparable integration error for every function in the associated reproducing kernel Hilbert space. With high probability, the maximum discrepancy in integration error is for compactly supported and for sub-exponential . In contrast, an equal-sized i.i.d. sample from suffers integration error. Our sub-exponential guarantees resemble the classical quasi-Monte Carlo error rates for uniform on but apply to general distributions on and a wide range of common kernels. We use our results to derive explicit non-asymptotic maximum mean discrepancy bounds for Gaussian, Mat\'ern, and B-spline kernels and present two vignettes illustrating the practical benefits of kernel thinning over i.i.d. sampling and standard Markov chain Monte Carlo thinning.
View on arXiv