CoSToM:Causal-oriented Steering for Intrinsic Theory-of-Mind Alignment in Large Language Models

arXiv cs.CL / 4/14/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that current LLM ToM performance often depends on prompt scaffolding and may not generalize to complex, task-specific scenarios, suggesting a mismatch between internal knowledge and external behavior.
  • It introduces CoSToM (Causal-oriented Steering for ToM alignment), which combines causal tracing to identify how ToM semantics are represented inside the model with targeted activation steering to intervene directly in ToM-critical layers.
  • By mapping the internal feature distributions through causal tracing, the method aims to shift from purely mechanistic interpretation toward active, behavior-stabilizing alignment.
  • Experiments reported in the paper indicate that CoSToM improves human-like social reasoning and enhances downstream dialogue quality.
  • Overall, the work proposes an approach for “intrinsic cognition” alignment by stabilizing externally observable ToM-like behavior through causal, internal interventions.

Abstract

Theory of Mind (ToM), the ability to attribute mental states to others, is a hallmark of social intelligence. While large language models (LLMs) demonstrate promising performance on standard ToM benchmarks, we observe that they often fail to generalize to complex task-specific scenarios, relying heavily on prompt scaffolding to mimic reasoning. The critical misalignment between the internal knowledge and external behavior raises a fundamental question: Do LLMs truly possess intrinsic cognition, and can they externalize this internal knowledge into stable, high-quality behaviors? To answer this, we introduce CoSToM (Causal-oriented Steering for ToM alignment), a framework that transitions from mechanistic interpretation to active intervention. First, we employ causal tracing to map the internal distribution of ToM features, empirically uncovering the internal layers' characteristics in encoding fundamental ToM semantics. Building on this insight, we implement a lightweight alignment framework via targeted activation steering within these ToM-critical layers. Experiments demonstrate that CoSToM significantly enhances human-like social reasoning capabilities and downstream dialogue quality.