Emergent Mind

Convergence of Graph Laplacian with kNN Self-tuned Kernels

(2011.01479)
Published Nov 3, 2020 in math.ST , cs.LG , stat.ML , and stat.TH

Abstract

Kernelized Gram matrix $W$ constructed from data points ${xi}{i=1}N$ as $W{ij}= k0( \frac{ | xi - xj |2} {\sigma2} )$ is widely used in graph-based geometric data analysis and unsupervised learning. An important question is how to choose the kernel bandwidth $\sigma$, and a common practice called self-tuned kernel adaptively sets a $\sigmai$ at each point $xi$ by the $k$-nearest neighbor (kNN) distance. When $xi$'s are sampled from a $d$-dimensional manifold embedded in a possibly high-dimensional space, unlike with fixed-bandwidth kernels, theoretical results of graph Laplacian convergence with self-tuned kernels have been incomplete. This paper proves the convergence of graph Laplacian operator $LN$ to manifold (weighted-)Laplacian for a new family of kNN self-tuned kernels $W{(\alpha)}_{ij} = k0( \frac{ | xi - xj |2}{ \epsilon \hat{\rho}(xi) \hat{\rho}(xj)})/\hat{\rho}(xi)\alpha \hat{\rho}(xj)\alpha$, where $\hat{\rho}$ is the estimated bandwidth function {by kNN}, and the limiting operator is also parametrized by $\alpha$. When $\alpha = 1$, the limiting operator is the weighted manifold Laplacian $\Deltap$. Specifically, we prove the point-wise convergence of $L_N f $ and convergence of the graph Dirichlet form with rates. Our analysis is based on first establishing a $C0$ consistency for $\hat{\rho}$ which bounds the relative estimation error $|\hat{\rho} - \bar{\rho}|/\bar{\rho}$ uniformly with high probability, where $\bar{\rho} = p{-1/d}$, and $p$ is the data density function. Our theoretical results reveal the advantage of self-tuned kernel over fixed-bandwidth kernel via smaller variance error in low-density regions. In the algorithm, no prior knowledge of $d$ or data density is needed. The theoretical results are supported by numerical experiments on simulated data and hand-written digit image data.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.