Guaranteeing Knowledge Integration with Joint Decoding for Retrieval-Augmented Generation
arXiv cs.CL / 4/10/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper identifies an “integration bottleneck” in retrieval-augmented generation (RAG), where LLMs may retrieve relevant documents but still fail to use them due to conflicts with their parametric knowledge.
- It proposes GuarantRAG, which decouples reasoning from evidence integration by generating an Inner-Answer from parametric knowledge and a Refer-Answer constrained by retrieved documents using a novel Contrastive DPO objective.
- A joint decoding mechanism then fuses the Inner-Answer’s logical coherence with the Refer-Answer’s factual precision at the token level, rather than relying on naive concatenation.
- Experiments across five QA benchmarks show improvements of up to 12.1% in accuracy and reductions in hallucinations by 16.3% versus standard and dynamic RAG baselines.
- Overall, the work frames evidence integration as an explicit, train-and-decode problem rather than a retrieval-quality problem alone.
Related Articles

Black Hat Asia
AI Business

GLM 5.1 tops the code arena rankings for open models
Reddit r/LocalLLaMA
Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to
My Bestie Built a Free MCP Server for Job Search — Here's How It Works
Dev.to
can we talk about how AI has gotten really good at lying to you?
Reddit r/artificial