Shadow-Loom: Causal Reasoning over Graphical World Model of Narratives

arXiv cs.AI / 5/5/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The Shadow-Loom framework converts a narrative into a versioned graphical world model that can be acted on by two reasoning engines.
  • One engine performs causal reasoning using Pearl’s ladder of causation, while the other uses a counterfactual calculus built for Ancestral Multi-World Networks.
  • The framework also introduces a “narrative physics” component that scores the same graph across four reader-oriented structural states (mystery, dramatic irony, suspense, and surprise).
  • Large language models are used only at the boundary for tasks like extraction, rendering, and auditing, while core identification, intervention, and counterfactual reasoning are executed in typed code over the graph.
  • The authors release Shadow-Loom as an experimental open-source research artifact with code, fixtures, and a pipeline, explicitly avoiding positioning it as a benchmark NLP model.

Abstract

Stories hold a reader's attention because they have causes, secrets, and consequences. Shadow-Loom is an experimental open-source framework that turns a narrative into a versioned graphical world model and lets two engines act on it: a causal physics grounded in Pearl's ladder of causation and a recently proposed counterfactual calculus over Ancestral Multi-World Networks; and a narrative physics that scores the same graph against four structural reader-states -- mystery, dramatic irony, suspense, and surprise -- in the tradition of Sternberg's curiosity/suspense/surprise triad, with suspense formalised in the structural-affect line of work on story comprehension and computational suspense. Large language models are used only at the boundary: extraction, rendering, and audit; identification, intervention, and counterfactual reasoning are carried out in typed code over the graph. The system is offered as a research artefact rather than as a benchmarked NLP model; code, fixtures, and pipeline are released open source.