When Do Hallucinations Arise? A Graph Perspective on the Evolution of Path Reuse and Path Compression

arXiv cs.AI / 4/7/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper analyzes reasoning hallucinations in decoder-only Transformers by reframing next-token prediction as graph search over an underlying learned entity-transition graph.
  • It distinguishes two modes of reasoning: contextual reasoning as constrained search in a sampled subgraph, versus context-free reasoning as reliance on memorized structures in the underlying graph.
  • The authors identify two core mechanisms behind hallucinations: Path Reuse (memorized knowledge overriding contextual constraints early in training) and Path Compression (frequent multi-step paths collapsing into shortcut edges later in training).
  • By unifying these mechanisms, the work offers an explanation for why hallucinations can be fluent yet inconsistent with both provided context and factual knowledge.
  • The findings connect the proposed graph-theoretic training dynamics to behaviors reported in downstream LLM applications, suggesting broader relevance beyond the specific modeling framework.

Abstract

Reasoning hallucinations in large language models (LLMs) often appear as fluent yet unsupported conclusions that violate either the given context or underlying factual knowledge. Although such failures are widely observed, the mechanisms by which decoder-only Transformers produce them remain poorly understood. We model next-token prediction as a graph search process over an underlying graph, where entities correspond to nodes and learned transitions form edges. From this perspective, contextual reasoning is a constrained search over a sampled subgraph (intrinsic reasoning), while context-free queries rely on memorized structures in the underlying graph (extrinsic reasoning). We show that reasoning hallucinations arise from two fundamental mechanisms: \textbf{Path Reuse}, where memorized knowledge overrides contextual constraints during early training, and \textbf{Path Compression}, where frequently traversed multi-step paths collapse into shortcut edges in later training. Together, these mechanisms provide a unified explanation for reasoning hallucinations in LLMs and connected to well-known behaviors observed in downstream applications.