Faithfulness-QA: A Counterfactual Entity Substitution Dataset for Training Context-Faithful RAG Models
arXiv cs.CL / 4/29/2026
📰 NewsTools & Practical UsageModels & Research
Key Points
- The paper targets an important RAG failure mode where models answer from parametric memory instead of the retrieved context, breaking the promise of retrieval grounding.
- It introduces Faithfulness-QA, a dataset with 99,094 counterfactually generated QA samples meant to force the model to prefer the provided context over internal knowledge.
- The dataset is built by counterfactual entity substitution: named entities in SQuAD and TriviaQA contexts are replaced with type-consistent alternatives from a curated bank of 76,953 entities, creating controlled conflicts.
- The authors apply rigorous quality filtering, reporting 100% pass rates across four automated checks on random audit samples, and release the dataset, pipeline, and typed entity bank for training and evaluation.
- Faithfulness-QA is intended both as training data for context-faithfulness objectives (e.g., attention-based) and as a benchmark to measure context-grounding behavior in RAG systems.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.


