Counterfactual Peptide Editing for Causal TCR--pMHC Binding Inference
arXiv cs.LG / 4/16/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- TCR–pMHC binding prediction models can suffer from shortcut learning, relying on spurious dataset correlations rather than the physical binding interface, which hurts generalization in tougher evaluation settings.
- The paper proposes Counterfactual Invariant Prediction (CIP), which creates biologically constrained counterfactual peptide edits and trains models to remain invariant to changes at non-anchor positions while becoming sensitive to disruptions at MHC anchor residues.
- CIP improves out-of-distribution performance on a curated VDJdb-IEDB benchmark, reaching an AUROC of 0.831 and a counterfactual consistency (CFC) of 0.724 under a family-held-out protocol.
- Compared with an unconstrained baseline, CIP reduces the shortcut index by 39.7%, and ablation results indicate that anchor-aware edit generation is the key driver of the OOD gains.
- The authors frame CIP as a practical recipe for causally grounded TCR specificity modeling rather than purely correlation-based prediction.
Related Articles

Black Hat Asia
AI Business
The AI Hype Cycle Is Lying to You About What to Learn
Dev.to
Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to
OpenAI Codex April 2026 Update Review: Computer Use, Memory & 90+ Plugins — Is the Hype Real?
Dev.to
Factory hits $1.5B valuation to build AI coding for enterprises
TechCrunch