What does a system modify when it modifies itself?
arXiv cs.AI / 3/31/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper asks what a cognitive system changes when it self-modifies—whether it updates low-level rules, control rules, or the evaluative norms behind its revisions—and argues that both cognitive science and contemporary AI lack a shared formal framework to distinguish these targets.
- It proposes a minimal formal structure for self-modifying systems—rule hierarchies, a fixed core, and a separation between effective, represented, and causally accessible rules—and derives four self-modification regimes (action-only, low-level modification, structural modification, and teleological revision).
- Applying the framework to humans, the authors claim a “crossing of opacities”: causal power and self-representation concentrate at higher hierarchical levels, while lower operational levels remain comparatively opaque.
- For reflexive AI systems, the paper argues the inverse pattern: operational levels have richer representation and causal accessibility, whereas the highest evaluative level lacks such access.
- The framework is linked to theories of artificial consciousness, yields four testable predictions, and lists four open problems including independence of transformativity vs. autonomy and identity under transformation.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat Asia
AI Business
[D] How does distributed proof of work computing handle the coordination needs of neural network training?
Reddit r/MachineLearning

Claude Code's Entire Source Code Was Just Leaked via npm Source Maps — Here's What's Inside
Dev.to

BYOK is not just a pricing model: why it changes AI product trust
Dev.to

AI Citation Registries and Identity Persistence Across Records
Dev.to