ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2011.01479
64
24
v1v2 (latest)

Convergence of Graph Laplacian with kNN Self-tuned Kernels

3 November 2020
Xiuyuan Cheng
Hau‐Tieng Wu
ArXiv (abs)PDFHTML
Abstract

Kernelized Gram matrix WWW constructed from data points {xi}i=1N\{x_i\}_{i=1}^N{xi​}i=1N​ as Wij=k0(∥xi−xj∥2σ2)W_{ij}= k_0( \frac{ \| x_i - x_j \|^2} {\sigma^2} ) Wij​=k0​(σ2∥xi​−xj​∥2​) is widely used in graph-based geometric data analysis and unsupervised learning. An important question is how to choose the kernel bandwidth σ\sigmaσ, and a common practice called self-tuned kernel adaptively sets a σi\sigma_iσi​ at each point xix_ixi​ by the kkk-nearest neighbor (kNN) distance. When xix_ixi​'s are sampled from a ddd-dimensional manifold embedded in a possibly high-dimensional space, unlike with fixed-bandwidth kernels, theoretical results of graph Laplacian convergence with self-tuned kernels, however, have been incomplete. This paper proves the convergence of graph Laplacian operator LNL_NLN​ to manifold (weighted-)Laplacian for a new family of kNN self-tuned kernels Wij(α)=k0(∥xi−xj∥2ϵρ^(xi)ρ^(xj))/ρ^(xi)αρ^(xj)αW^{(\alpha)}_{ij} = k_0( \frac{ \| x_i - x_j \|^2}{ \epsilon \hat{\rho}(x_i) \hat{\rho}(x_j)})/\hat{\rho}(x_i)^\alpha \hat{\rho}(x_j)^\alphaWij(α)​=k0​(ϵρ^​(xi​)ρ^​(xj​)∥xi​−xj​∥2​)/ρ^​(xi​)αρ^​(xj​)α, where ρ^\hat{\rho}ρ^​ is the estimated bandwidth function {by kNN}, and the limiting operator is also parametrized by α\alphaα. When α=1\alpha = 1α=1, the limiting operator is the weighted manifold Laplacian Δp\Delta_pΔp​. Specifically, we prove the pointwise convergence of LNfL_N f LN​f and convergence of the graph Dirichlet form with rates. Our analysis is based on first establishing a C0C^0C0 consistency for ρ^\hat{\rho}ρ^​ which bounds the relative estimation error ∣ρ^−ρˉ∣/ρˉ|\hat{\rho} - \bar{\rho}|/\bar{\rho}∣ρ^​−ρˉ​∣/ρˉ​ uniformly with high probability, where ρˉ=p−1/d\bar{\rho} = p^{-1/d}ρˉ​=p−1/d, and ppp is the data density function. Our theoretical results reveal the advantage of self-tuned kernel over fixed-bandwidth kernel via smaller variance error in low-density regions. In the algorithm, no prior knowledge of ddd or data density is needed. The theoretical results are supported by numerical experiments.

View on arXiv
Comments on this paper