AI Navigate

Open-Source Reproduction and Explainability Analysis of Corrective Retrieval Augmented Generation

arXiv cs.CL / 3/18/2026

💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The paper presents a fully open-source reproduction of CRAG, replacing proprietary web search with the Wikipedia API and replacing the LLaMA-2 generator with Phi-3-mini-4k-instruct to improve reproducibility.
  • It evaluates on PopQA and ARC-Challenge, showing the open-source pipeline achieves comparable performance to the original CRAG system.
  • The work includes the first explainability analysis of CRAG's T5-based retrieval evaluator using SHAP, revealing reliance on named entity alignment rather than semantic similarity.
  • The study identifies key failure modes such as domain transfer limitations on science questions and provides the code and results at the linked GitHub repository.

Abstract

Corrective Retrieval Augmented Generation (CRAG) improves the robustness of RAG systems by evaluating retrieved document quality and triggering corrective actions. However, the original implementation relies on proprietary components including the Google Search API and closed model weights, limiting reproducibility. In this work, we present a fully open-source reproduction of CRAG, replacing proprietary web search with the Wikipedia API and the original LLaMA-2 generator with Phi-3-mini-4k-instruct. We evaluate on PopQA and ARC-Challenge, demonstrating that our open-source pipeline achieves comparable performance to the original system. Furthermore, we contribute the first explainability analysis of CRAG's T5-based retrieval evaluator using SHAP, revealing that the evaluator primarily relies on named entity alignment rather than semantic similarity. Our analysis identifies key failure modes including domain transfer limitations on science questions. All code and results are available at https://github.com/suryayalavarthi/crag-reproduction.