Sinkhorn Based Associative Memory Retrieval Using Spherical Hellinger Kantorovich Dynamics

arXiv stat.ML / 3/24/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a dense associative memory framework where both stored patterns and retrieval queries are represented as finitely supported probability measures (weighted point clouds).
  • Retrieval is formulated as minimizing a Hopfield-style log-sum-exp energy based on a debiased Sinkhorn divergence, defining a probabilistic analogue of energy minimization for measure-valued data.
  • The authors derive retrieval dynamics as a spherical Hellinger Kantorovich (SHK) gradient flow that simultaneously updates both the locations (support) and weights of the retrieved measure.
  • A discretized, deterministic algorithm is proposed that uses Sinkhorn potentials for barycentric transport steps along with multiplicative simplex reweighting to implement the gradient flow numerically.
  • Theoretical results include basin invariance and geometric convergence under local separation/PL-type assumptions, plus a random pattern model showing disjoint Sinkhorn basins with high probability and thus exponential capacity in dimension; experiments on synthetic Gaussian memories show robust recovery versus a Euclidean Hopfield baseline.

Abstract

We propose a dense associative memory for empirical measures (weighted point clouds). Stored patterns and queries are finitely supported probability measures, and retrieval is defined by minimizing a Hopfield-style log-sum-exp energy built from the debiased Sinkhorn divergence. We derive retrieval dynamics as a spherical Hellinger Kantorovich (SHK) gradient flow, which updates both support locations and weights. Discretizing the flow yields a deterministic algorithm that uses Sinkhorn potentials to compute barycentric transport steps and a multiplicative simplex reweighting. Under local separation and PL-type conditions we prove basin invariance, geometric convergence to a local minimizer, and a bound showing the minimizer remains close to the corresponding stored pattern. Under a random pattern model, we further show that these Sinkhorn basins are disjoint with high probability, implying exponential capacity in the ambient dimension. Experiments on synthetic Gaussian point-cloud memories demonstrate robust recovery from perturbed queries versus a Euclidean Hopfield-type baseline.