Global and Local Topology-Aware Attention with Persistent Homology and Euler Biases for Time-Series Forecasting

arXiv cs.LG / 5/6/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a topology-aware attention framework for time-series forecasting that injects geometric and connectivity structure into attention logits using persistent homology (H0–H2), anchored Euler characteristic transforms, and kernel-Hilbert channels.
  • It introduces validation-gated local residual mechanisms that apply local topological signals only when held-out validation data supports the correction, using exact Vietoris–Rips computations and smooth topological surrogates under a no-leakage evaluation protocol.
  • Experiments across three architecture families (lightweight attention/Ridge, PatchTST, and a TimeSeriesTransformer) on synthetic benchmarks plus real datasets (CO2, S&P 500 return-window geometry, NASA IMS bearing degradation) show consistent positive paired effects when topology is predictive.
  • Reported performance gains include mean relative RMSE reductions of about 12.5% (lightweight attention/Ridge), 23.5% (PatchTST), and 47.8% (TimeSeriesTransformer), with strong statistical significance from matched paired comparisons.
  • Overall, the authors argue that topology can function as a validation-selected, architecture-compatible inductive bias for forecasting tasks where underlying geometry is informative.

Abstract

Scientific time series often encode predictive geometric structure, including connectivity, cycles, shell-like geometry, directional changes, and nonlinear neighborhoods, that standard dot-product attention does not explicitly represent. We introduce a topology-aware attention framework that adds such structure to attention logits using persistent homology (H0-H2), anchored Euler characteristic transforms, and kernel-Hilbert channels. A validation-gated local residual captures local topological signals, including a Zeng-style local H0 component, only when held-out validation data support the correction. Exact Vietoris-Rips computations and smooth topological surrogates are evaluated under a no-leakage protocol with train-only calibration, validation-only selection, and test-only reporting. We evaluate guarded topology-aware variants across three architecture families: lightweight attention/Ridge, PatchTSTForRegression, and TimeSeriesTransformerForPrediction. Experiments include synthetic benchmarks isolating higher-order topology and real datasets covering CO2, S&P 500 return-window geometry, and NASA IMS bearing degradation. The audit uses matched paired comparisons across seven dataset units, three random seeds, and three chronological splits, giving 63 paired units per architecture and 189 paired units overall. Topology-aware models show positive paired effects when geometry is predictive, with heterogeneous magnitude across datasets and architectures. Lightweight attention/Ridge improves in 46 of 63 units, with mean relative RMSE reduction of 12.5% and paired randomization p=7.2e-4; PatchTST improves in 33 units and retains the baseline in 20 units, with 23.5% reduction and p=3.5e-5; and TimeSeriesTransformer improves in 47 units, with 47.8% reduction and p<1e-4. The results support topology as a validation-selected, architecture-compatible inductive bias.