Quantifying Cross-Query Contradictions in Multi-Query LLM Reasoning
arXiv cs.AI / 4/17/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper studies why LLMs often generate mutually inconsistent answers across multiple related queries and frames it as maintaining a globally satisfiable belief state.
- It introduces a new benchmark of 390 multi-query reasoning instances labeled as entailment, contradiction, or unknown, along with set-level evaluation metrics such as Case Satisfiability Rate, Contradiction Density, and Revision Cost.
- A solver-augmented method is proposed that extracts model commitments, checks global satisfiability, and uses counterexample-guided repair to fix inconsistencies.
- Experiments across four reasoning domains show the approach substantially reduces cross-query contradictions (SetCons: 0.56 to 0.94) without sacrificing per-query accuracy, highlighting the importance of global coherence.
Related Articles
langchain-anthropic==1.4.1
LangChain Releases

🚀 Anti-Gravity Meets Cloud AI: The Future of Effortless Development
Dev.to

Talk to Your Favorite Game Characters! Mantella Brings AI to Skyrim and Fallout 4 NPCs
Dev.to

AI Will Run Companies. Here's Why That Should Excite You, Not Scare You.
Dev.to

The problem with Big Tech AI pricing (and why 8 countries can't afford to compete)
Dev.to