Monitoring Neural Training with Topology: A Footprint-Predictable Collapse Index

arXiv cs.LG / 5/1/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper introduces an online monitor to detect representational collapse early by tracking changes in neural embeddings before standard performance metrics decline.
  • It combines Modular Morse Homology Maintenance (MMHM) with a composite Collapse Index (CI) to assess anisotropy and loss of multi-scale structure in evolving representations.
  • The method avoids expensive full recomputation each epoch by using sparse, fixed-scale edits and maintaining a discrete Morse matching for fast incremental updates.
  • Experiments on LLM fine-tuning and temporal knowledge graph embedding (KGE) training show that CI can serve as a low-latency early-warning signal that supports in-training interventions.
  • The authors plan to publicly release the code and experimental scripts.

Abstract

Representational collapse, where embeddings become anisotropic and lose multi-scale structure, can erode downstream performance long before performance metrics react. We propose an online, topology-aware monitor for evolving neural representations that couples Modular Morse Homology Maintenance (MMHM) with a composite Collapse Index (CI). Instead of rebuilding complexes each epoch, we apply sparse edits at a fixed scale and maintain a discrete Morse matching, yielding fast, incremental updates. Across LLM fine-tuning and temporal KGE training, CI provides a low-latency early-warning signal suitable for in-training interventions. Code and experimental scripts will be released publicly