Epistemic Compression: The Case for Deliberate Ignorance in High-Stakes AI

arXiv cs.LG / 3/27/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that foundation models often underperform in high-stakes, reliability-critical domains (medicine, finance, policy) due to a “Fidelity Paradox” that is structural rather than purely a data issue.
  • It introduces “Epistemic Compression,” claiming robustness comes from aligning model complexity with the effective shelf life (stability) of the training data rather than from simply scaling parameters.
  • The method differs from classical regularization by enforcing parsimony at the architectural level, making it inherently costly for the model to encode variance not supported by the evidence in the data.
  • It operationalizes the approach using a “Regime Index” that distinguishes between Shifting Regimes (unstable, data-poor—favor simplicity) and Stable Regimes (invariant, data-rich—allow complexity).
  • In an exploratory synthesis across 15 high-stakes domains, the Regime Index matched the empirically better strategy in 86.7% of cases (13/15), supporting the proposed shift toward principled parsimony for high-stakes AI.

Abstract

Foundation models excel in stable environments, yet often fail where reliability matters most: medicine, finance, and policy. This Fidelity Paradox is not just a data problem; it is structural. In domains where rules change over time, extra model capacity amplifies noise rather than capturing signal. We introduce Epistemic Compression: the principle that robustness emerges from matching model complexity to the shelf life of the data, not from scaling parameters. Unlike classical regularization, which penalizes weights post hoc, Epistemic Compression enforces parsimony through architecture: the model structure itself is designed to reduce overfitting by making it architecturally costly to represent variance that exceeds the evidence in the data. We operationalize this with a Regime Index that separates Shifting Regime (unstable, data-poor; simplicity wins) from Stable Regime (invariant, data-rich; complexity viable). In an exploratory synthesis of 15 high-stakes domains, this index was concordant with the empirically superior modeling strategy in 86.7% of cases (13/15). High-stakes AI demands a shift from scaling for its own sake to principled parsimony.