We develop two approaches for analyzing the approximation error bound for the Nystr\"{o}m method, one based on the concentration inequality of integral operator, and one based on the compressive sensing theory. We show that the approximation error, measured in the spectral norm, can be improved from to in the case of large eigengap, where is the total number of data points, is the number of sampled data points, and is a positive constant that characterizes the eigengap. When the eigenvalues of the kernel matrix follow a -power law, our analysis based on compressive sensing theory further improves the bound to under an incoherence assumption, which explains why the Nystr\"{o}m method works well for kernel matrix with skewed eigenvalues. We present a kernel classification approach based on the Nystr\"{o}m method and derive its generalization performance using the improved bound. We show that when the eigenvalues of kernel matrix follow a -power law, we can reduce the number of support vectors to , a number less than when , without seriously sacrificing its generalization performance.
View on arXiv