Adaptive Kernel Selection for Kernelized Diffusion Maps

arXiv stat.ML / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses kernel selection as a core challenge in kernel-based spectral methods, showing how the choice of kernel in Kernelized Diffusion Maps (KDM) affects RKHS estimation accuracy and the stability/quality of recovered eigenfunctions.
  • It proposes two complementary adaptive strategies: a variational outer loop that learns continuous kernel parameters via differentiating through a Cholesky-reduced KDM eigenproblem using an objective based on eigenvalue maximization, subspace orthonormality, and RKHS regularization.
  • It also introduces an unsupervised cross-validation pipeline that selects kernel families and bandwidths using an eigenvalue-sum criterion, enhanced with random Fourier features for scalability.
  • The work includes a unified theoretical foundation proving Lipschitz dependence of KDM operators on kernel weights, continuity of spectral projectors under a spectral-gap condition, a residual-control theorem guaranteeing closeness to the target eigenspace, and exponential consistency of the cross-validation selector over a finite kernel dictionary.

Abstract

Selecting an appropriate kernel is a central challenge in kernel-based spectral methods. In \emph{Kernelized Diffusion Maps} (KDM), the kernel determines the accuracy of the RKHS estimator of a diffusion-type operator and hence the quality and stability of the recovered eigenfunctions. We introduce two complementary approaches to adaptive kernel selection for KDM. First, we develop a variational outer loop that learns continuous kernel parameters, including bandwidths and mixture weights, by differentiating through the Cholesky-reduced KDM eigenproblem with an objective combining eigenvalue maximization, subspace orthonormality, and RKHS regularization. Second, we propose an unsupervised cross-validation pipeline that selects kernel families and bandwidths using an eigenvalue-sum criterion together with random Fourier features for scalability. Both methods share a common theoretical foundation: we prove Lipschitz dependence of KDM operators on kernel weights, continuity of spectral projectors under a gap condition, a residual-control theorem certifying proximity to the target eigenspace, and exponential consistency of the cross-validation selector over a finite kernel dictionary.

Adaptive Kernel Selection for Kernelized Diffusion Maps | AI Navigate