When Verification Hurts: Asymmetric Effects of Multi-Agent Feedback in Logic Proof Tutoring

arXiv cs.AI / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how LLM-based tutoring with step-level feedback for propositional logic proofs behaves under verification, focusing on correctness aligned to a learner’s current proof state.
  • It introduces a knowledge-graph-grounded benchmark of 516 annotated proof states, enabling fine-grained evaluation of feedback quality against verified solution paths.
  • Across three role-specialized multi-agent pipelines (Tutor with partial solution access, Teacher with full derivations, and Judge verifying Tutor feedback), the authors find an asymmetric effect: verification helps when upstream feedback is inaccurate but harms by 4–6 points when upstream feedback is already reliable.
  • The analysis attributes degradation to over-specification and reports a shared “complexity ceiling,” with no approach reliably solving proof states beyond complexity level 4–5.
  • The findings challenge the idea that adding verifiers or richer context always improves tutoring performance, suggesting the need for adaptive, difficulty-aware routing based on estimated complexity and upstream reliability.

Abstract

Large language models (LLMs) are increasingly used for automated tutoring, but their reliability in structured symbolic domains remains unclear. We study step-level feedback for propositional logic proofs, which require precise symbolic reasoning aligned with a learner's current proof state. We introduce a knowledge-graph-grounded benchmark of 516 unique proof states with step-level annotations and difficulty metrics. Unlike prior tutoring evaluations that rely on model self-assessment or binary correctness, our framework enables fine-grained analysis of feedback quality against verified solution paths. We evaluate three role-specialized pipelines with varying solution access: Tutor (partial solution access), Teacher (full derivation access), and Judge (verification of Tutor feedback). Our results reveal a striking asymmetry: verification improves outcomes when upstream feedback is error-prone (<70% accuracy), but degrades performance by 4-6 percentage points through over-specification when feedback is already reliable (>85%). Critically, we identify a shared complexity ceiling; no model or pipeline reliably succeeds on proof states exceeding complexity 4-5. These findings challenge the assumption that adding verifiers or richer context universally improves tutoring, motivating adaptive, difficulty-aware architectures that route problems by estimated complexity and upstream reliability.