Supervised Learning Has a Necessary Geometric Blind Spot: Theory, Consequences, and Minimal Repair

arXiv cs.LG / 4/24/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proves that empirical risk minimisation (ERM) forces a “geometric blind spot” on learned representations, meaning supervised-loss minimisation must keep label-correlated Jacobian sensitivity that becomes a nuisance at test time.
  • This geometric blind spot is shown to underlie multiple previously separate phenomena—non-robust predictive features, texture bias, corruption fragility, and the robustness–accuracy tradeoff—casting adversarial vulnerability as a structural consequence of supervised learning geometry.
  • The authors introduce the Trajectory Deviation Index (TDI) as a diagnostic metric that directly captures the quantity bounded by their theorem, explaining why alternatives can miss the key failure mode.
  • Experiments across multiple vision and language settings (including BERT/SST-2 and ImageNet ViT backbones used in CLIP/DINO/SAM) show the blind spot is measurable, worsens with scale in language models, and can be repaired using a method called PMH that reduces the issue by about 11× with an added Gaussian-regularisation training term.
  • The blind spot is not portrayed as a limitation of current architectures or datasets; it is claimed to hold across proper scoring rules, architectures, and dataset sizes, and remains even at foundation-model scale.

Abstract

We prove that empirical risk minimisation (ERM) imposes a necessary geometric constraint on learned representations: any encoder that minimises supervised loss must retain non-zero Jacobian sensitivity in directions that are label-correlated in training data but nuisance at test time. This is not a contingent failure of current methods; it is a mathematical consequence of the supervised objective itself. We call this the geometric blind spot of supervised learning (Theorem 1), and show it holds across proper scoring rules, architectures, and dataset sizes. This single theorem unifies four lines of prior empirical work that were previously treated separately: non-robust predictive features, texture bias, corruption fragility, and the robustness-accuracy tradeoff. In this framing, adversarial vulnerability is one consequence of a broader structural fact about supervised learning geometry. We introduce Trajectory Deviation Index (TDI), a diagnostic that measures the theorem's bounded quantity directly, and show why common alternatives miss the key failure mode. PGD adversarial training reaches Jacobian Frobenius 2.91 yet has the worst clean-input geometry (TDI 1.336), while PMH achieves TDI 0.904. TDI is the only metric that detects this dissociation because it measures isotropic path-length distortion -- the exact quantity Theorem 1 bounds. Across seven vision tasks, BERT/SST-2, and ImageNet ViT-B/16 backbones used by CLIP, DINO, and SAM, the blind spot is measurable and repairable. It is present at foundation-model scale, worsens monotonically across language-model sizes (blind-spot ratio 0.860 to 0.765 to 0.742 from 66M to 340M), and is amplified by task-specific ERM fine-tuning (+54%), while PMH repairs it by 11x with one additional training term whose Gaussian form Proposition 5 proves is the unique perturbation law that uniformly penalises the encoder Jacobian.