Epistemic Blinding: An Inference-Time Protocol for Auditing Prior Contamination in LLM-Assisted Analysis

arXiv cs.AI / 4/8/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The paper identifies “epistemic blinding” as a way to audit LLM-assisted analysis when model training priors silently blend with data provided in the prompt, making it impossible to tell which source drove a given output.
  • It proposes an inference-time protocol that replaces entity identifiers with anonymous codes, then compares results to an unblinded control to estimate the degree of prior contamination.
  • In an oncology drug-target prioritization system, blinding changes 16% of top-20 predictions while still recovering validated targets, suggesting improved auditability without sacrificing key findings.
  • The contamination issue is shown to generalize beyond biology: in S&P 500 equity screening, brand-recognition bias alters 30–40% of top-20 rankings across multiple runs.
  • To support adoption, the authors release an open-source tool and a Claude Code skill that enables one-command epistemic blinding in agentic workflows.

Abstract

This paper presents epistemic blinding in the context of an agentic system that uses large language models to reason across multiple biological datasets for drug target prioritization. During development, it became apparent that LLM outputs silently blend data-driven inference with memorized priors about named entities - and the blend is invisible: there is no way to determine, from a single output, how much came from the data on the page and how much came from the model's training memory. Epistemic blinding is a simple inference-time protocol that replaces entity identifiers with anonymous codes before prompting, then compares outputs against an unblinded control. The protocol does not make LLM reasoning deterministic, but it restores one critical axis of auditability: measuring how much of an output came from the supplied data versus the model's parametric knowledge. The complete target identification system is described - including LLM-guided evolutionary optimization of scoring functions and blinded agentic reasoning for target rationalization - with demonstration that both stages operate without access to entity identity. In oncology drug target prioritization across four cancer types, blinding changes 16% of top-20 predictions while preserving identical recovery of validated targets. The contamination problem is shown to generalize beyond biology: in S&P 500 equity screening, brand-recognition bias reshapes 30-40% of top-20 rankings across five random seeds. To lower the barrier to adoption, the protocol is released as an open-source tool and as a Claude Code skill that enables one-command epistemic blinding within agentic workflows. The claim is not that blinded analysis produces better results, but that without blinding, there is no way to know to what degree the agent is adhering to the analytical process the researcher designed.