Bridging the Know-Act Gap via Task-Level Autoregressive Reasoning

arXiv cs.AI / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that LLMs can recognize flawed or ill-posed inputs under discriminative prompting but still generate plausible-sounding answers in standard generation, creating a “know-act gap.”
  • It introduces FaultyScience, a new large-scale cross-disciplinary benchmark for faulty scientific questions, and finds the know-act gap is pervasive rather than limited to narrow QA/math settings.
  • The authors attribute the gap to token-level autoregression that entangles task selection (e.g., validate vs. answer) with content generation, preventing the model’s discriminative knowledge from being acted upon.
  • To bridge this, they propose DeIllusionLLM, a task-level autoregressive framework that explicitly models the decision between discriminative validation and generative answering.
  • Experiments report that self-distillation enables a single model backbone to combine discriminative judgment with generative reasoning, substantially reducing “answer-despite-error” failures under natural prompting while preserving overall reasoning performance.

Abstract

LLMs often generate seemingly valid answers to flawed or ill-posed inputs. This is not due to missing knowledge: under discriminative prompting, the same models can mostly identify such issues, yet fail to reflect this in standard generative responses. This reveals a fundamental know-act gap between discriminative recognition and generative behavior. Prior work largely characterizes this issue in narrow settings, such as math word problems or question answering, with limited focus on how to integrate these two modes. In this work, we present a comprehensive analysis using FaultyScience, a newly constructed large-scale, cross-disciplinary benchmark of faulty scientific questions. We show that the gap is pervasive and stems from token-level autoregression, which entangles task selection (validate vs. answer) with content generation, preventing discriminative knowledge from being utilized. To address this, we propose DeIllusionLLM, a task-level autoregressive framework that explicitly models this decision. Through self-distillation, the model unifies discriminative judgment and generative reasoning within a single backbone. Empirically, DeIllusionLLM substantially reduces answer-despite-error failures under natural prompting while maintaining general reasoning performance, demonstrating that self-distillation is an effective and scalable solution for bridging the discriminative-generative know-act gap