Beyond Theory of Mind in Robotics

arXiv cs.AI / 4/14/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper critiques the prevailing Theory of Mind (ToM) paradigm in robotics as being based on assumptions that do not reflect how real-world social interaction unfolds.
  • It argues that social meaning is produced through moment-to-moment coordination between agents, rather than inferred “inside-out” from hidden mental states by a detached observer.
  • Drawing on ethnomethodology, conversation analysis, and participatory sense-making, the work reframes understanding as participatory rather than purely inferential.
  • The authors propose design implications for robots: move from internal state modeling to coordination-sustaining policies, and from fixed behavioral meaning to meaning potential stabilized through responsive interaction.

Abstract

Theory of Mind, the capacity to explain and predict behavior by inferring hidden mental states, has become the dominant paradigm for social interaction in robotics. Yet ToM rests on three assumptions that poorly capture how most social interaction actually unfolds: that meaning travels inside-out from hidden states to observable behavior; that understanding requires detached inference rather than participation; and that the meaning of behavior is fixed and available to a passive observer. Drawing on ethnomethodology, conversation analysis, and participatory sense-making, I argue that social meaning is not decoded from behavior but produced through moment-to-moment coordination between agents. This interactional foundation has direct implications for robot design: shifting from internal state modeling toward policies for sustaining coordination, from observer-based inference toward active participation, and from fixed behavioral meaning toward meaning potential stabilized through response.