Reasoning Graphs: Deterministic Agent Accuracy through Evidence-Centric Chain-of-Thought Feedback
arXiv cs.CL / 4/10/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that agent “chain-of-thought” resets between similar queries, causing lower accuracy and high run-to-run variance because prior deliberations are discarded.
- It introduces reasoning graphs, which persist an agent’s deliberation tied to specific retrieved evidence by storing structured, evidence-connected edges that can be traversed in later runs.
- It contrasts this evidence-centric backward traversal with prior memory approaches that retrieve by query similarity or recency, emphasizing that feedback is tied to the currently evaluated evidence rather than the query.
- It also proposes retrieval graphs to iteratively tighten the candidate set via a pipeline planner, and claims that the combined graphs form a self-improving loop that boosts accuracy while collapsing variance without retraining.
- The authors formalize the structures and traversal algorithms and outline an evaluation protocol (sequential cluster evaluation) to measure accuracy convergence on multi-hop question answering benchmarks.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat Asia
AI Business

GLM 5.1 tops the code arena rankings for open models
Reddit r/LocalLLaMA

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

My Bestie Built a Free MCP Server for Job Search — Here's How It Works
Dev.to
can we talk about how AI has gotten really good at lying to you?
Reddit r/artificial