Evidence of an Emergent "Self" in Continual Robot Learning

arXiv cs.RO / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a quantitative framework for identifying an emergent “self” in intelligent systems by isolating the cognitive processes that remain invariant while other knowledge rapidly changes.
  • Using two continual-robot-learning setups, it finds that robots exposed to variable tasks develop an invariant subnetwork that is statistically more stable than a robot trained on a constant task (p < 0.001).
  • The authors interpret this stability as evidence consistent with a persistent internal “self”-like structure emerging from continual learning dynamics.
  • They argue the same invariance-based principle could be used to study selfhood in other cognitive AI systems beyond robots.
  • The work is positioned as a conceptual bridge between self-awareness theory and measurable neural/cognitive structure in learning agents.

Abstract

A key challenge to understanding self-awareness has been a principled way of quantifying whether an intelligent system has a concept of a "self," and if so how to differentiate the "self" from other cognitive structures. We propose that the "self" can be isolated by seeking the invariant portion of cognitive process that changes relatively little compared to more rapidly acquired cognitive knowledge and skills, because our self is the most persistent aspect of our experiences. We used this principle to analyze the cognitive structure of robots under two conditions: One robot learns a constant task, while a second robot is subjected to continual learning under variable tasks. We find that robots subjected to continual learning develop an invariant subnetwork that is significantly more stable (p < 0.001) compared to the control. We suggest that this principle can offer a window into exploring selfhood in other cognitive AI systems.