FACT-E: Causality-Inspired Evaluation for Trustworthy Chain-of-Thought Reasoning

arXiv cs.AI / 4/14/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • Chain-of-Thought (CoT) can look convincing while using unfaithful intermediate steps, making existing self-evaluation methods unreliable due to coherence-bias effects.
  • FACT-E introduces a causality-inspired evaluation approach using controlled perturbations to more reliably measure intra-chain faithfulness (true step-to-step dependence).
  • The method selects more trustworthy reasoning trajectories by jointly optimizing intra-chain faithfulness and CoT-to-answer consistency.
  • Experiments on GSM8K, MATH, and CommonsenseQA show FACT-E improves the selection of reasoning trajectories and strengthens in-context learning exemplars.
  • FACT-E also demonstrates robustness by detecting flawed reasoning more reliably under noisy conditions, offering a sturdier metric for trustworthy LLM reasoning.

Abstract

Chain-of-Thought (CoT) prompting has improved LLM reasoning, but models often generate explanations that appear coherent while containing unfaithful intermediate steps. Existing self-evaluation approaches are prone to inherent biases: the model may confidently endorse coherence even when the step-to-step implication is not valid, leading to unreliable faithfulness evaluation. We propose FACT-E, a causality-inspired framework for evaluating CoT quality. FACT-E uses controlled perturbations as an instrumental signal to separate genuine step-to-step dependence from bias-driven artifacts, producing more reliable faithfulness estimates (\textit{intra-chain faithfulness}). To select trustworthy trajectories, FACT-E jointly considers \textit{intra-chain faithfulness} and \textit{CoT-to-answer consistency}, ensuring that selected chains are both faithful internally and supportive of the correct final answer. Experiments on GSM8K, MATH, and CommonsenseQA show that FACT-E improves reasoning-trajectory selection and yields stronger in-context learning exemplars. FACT-E also reliably detects flawed reasoning under noisy conditions, providing a robust metric for trustworthy LLM reasoning.