Hallucinations Undermine Trust; Metacognition is a Way Forward

arXiv cs.CL / 5/5/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • Even with improved factual reliability, generative AI (especially LLMs) still produces hallucinations, including in simple factoid QA where ground truth is clear and no external tools are used.
  • The paper argues that many past gains have come from widening the model’s knowledge (encoding more facts), not from better awareness of what the model knows versus does not know.
  • It suggests a fundamental difficulty: models may not have sufficient discriminative power to perfectly separate truths from errors, creating a tradeoff between reducing hallucinations and maintaining usefulness.
  • As an alternative to the answer-or-abstain choice, the authors propose “faithful uncertainty,” where linguistic uncertainty should reflect intrinsic uncertainty, tying this to metacognition.
  • They conclude that metacognition—knowing and acting on one’s own uncertainty—is critical for making LLMs both trustworthy in direct chats and controllable in agentic systems (e.g., deciding when to search and what to trust).

Abstract

Despite significant strides in factual reliability, errors -- often termed hallucinations -- remain a major concern for generative AI, especially as LLMs are increasingly expected to be helpful in more complex or nuanced setups. Yet even in the simplest setting -- factoid question-answering with clear ground truth-frontier models without external tools continue to hallucinate. We argue that most factuality gains in this domain have come from expanding the model's knowledge boundary (encoding more facts) rather than improving awareness of that boundary (distinguishing known from unknown). We conjecture that the latter is inherently difficult: models may lack the discriminative power to perfectly separate truths from errors, creating an unavoidable tradeoff between eliminating hallucinations and preserving utility. This tradeoff dissolves under a different framing. If we understand hallucinations as confident errors -- incorrect information delivered without appropriate qualification -- a third path emerges beyond the answer-or-abstain dichotomy: expressing uncertainty. We propose faithful uncertainty: aligning linguistic uncertainty with intrinsic uncertainty. This is one facet of metacognition -- the ability to be aware of one's own uncertainty and to act on it. For direct interaction, acting on uncertainty means communicating it honestly; for agentic systems, it becomes the control layer governing when to search and what to trust. Metacognition is thus essential for LLMs to be both trustworthy and capable; we conclude by highlighting open problems for progress towards this objective.