316

Convergence of Graph Laplacian with kNN Self-tuned Kernels

Abstract

Kernelized Gram matrix WW constructed from data points {xi}i=1N\{x_i\}_{i=1}^N as $W_{ij}= k_0( \frac{ \| x_i - x_j \|^2} {\sigma^2} ) $ is widely used in graph-based geometric data analysis and unsupervised learning. An important question is how to choose the kernel bandwidth σ\sigma, and a common practice called self-tuned kernel adaptively sets a σi\sigma_i at each point xix_i by the kk-nearest neighbor (kNN) distance. When xix_i's are sampled from a dd-dimensional manifold embedded in a possibly high-dimensional space, unlike with fixed-bandwidth kernels, theoretical results of graph Laplacian convergence with self-tuned kernels, however, have been incomplete. This paper proves the convergence of graph Laplacian operator LNL_N to manifold (weighted-)Laplacian for a new family of kNN self-tuned kernels Wij(α)=k0(xixj2ϵρ^(xi)ρ^(xj))/ρ^(xi)αρ^(xj)αW^{(\alpha)}_{ij} = k_0( \frac{ \| x_i - x_j \|^2}{ \epsilon \hat{\rho}(x_i) \hat{\rho}(x_j)})/\hat{\rho}(x_i)^\alpha \hat{\rho}(x_j)^\alpha, where ρ^\hat{\rho} is the estimated bandwidth function {by kNN}, and the limiting operator is also parametrized by α\alpha. When α=1\alpha = 1, the limiting operator is the weighted manifold Laplacian Δp\Delta_p. Specifically, we prove the pointwise convergence of $L_N f $ and convergence of the graph Dirichlet form with rates. Our analysis is based on first establishing a C0C^0 consistency for ρ^\hat{\rho} which bounds the relative estimation error ρ^ρˉ/ρˉ|\hat{\rho} - \bar{\rho}|/\bar{\rho} uniformly with high probability, where ρˉ=p1/d\bar{\rho} = p^{-1/d}, and pp is the data density function. Our theoretical results reveal the advantage of self-tuned kernel over fixed-bandwidth kernel via smaller variance error in low-density regions. In the algorithm, no prior knowledge of dd or data density is needed. The theoretical results are supported by numerical experiments.

View on arXiv
Comments on this paper