Diffusion Operator Geometry of Feedforward Representations

arXiv cs.LG / 5/5/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a smooth, operator-theoretic framework to analyze the geometry of feedforward neural network representations, avoiding reliance on discrete neighborhood-graph curvature.
  • It models each feature cloud as inducing a Gaussian-kernel diffusion Markov operator and derives multiple observables (transport, spectral properties, label-boundary behavior, and local scales) using Bakry–Émery Γ-calculus.
  • In a balanced Gaussian class-conditional snapshot model with shared covariance, the population operator yields closed-form quantities such as class affinities, leakage, and coarse spectral structure, governed by pairwise regularized Mahalanobis separations.
  • The authors show these operator-based observables change smoothly under feature perturbations, whereas traditional neighborhood-graph diagnostics can shift discontinuously.
  • Experiments on synthetic Gaussian data and on learned MNIST demonstrate that the derived operator observables track training dynamics, network width, and perturbation stability.

Abstract

Neural networks transform data through learned representations whose geometry affects separation, contraction, and generalization. Recent work studies this geometry using discrete curvature on neighborhood graphs, suggesting Ricci-flow-like behavior across layers. We develop a smooth operator-theoretic alternative for feedforward representation snapshots. Each feature cloud induces a Gaussian-kernel diffusion Markov operator, and transport, spectral, label-boundary, and local-scale observables are derived from this single object via Bakry-Emery \Gamma-calculus. In a balanced Gaussian class-conditional snapshot model with shared covariance, the population operator has closed-form class affinities, leakage, and coarse spectra, all controlled by pairwise regularized Mahalanobis separations c_\varepsilon^{(a,b)}. We also prove that the resulting operator observables vary smoothly under feature perturbations, while hard neighborhood-graph diagnostics can change discontinuously. Synthetic experiments validate the closed-form Gaussian bridge, while learned MNIST experiments show that the same operator observables track training, width, and perturbation stability. Together, these results give a stable operator-geometric framework for analyzing feedforward representation geometry.