AI Safety as Control of Irreversibility: A Systems Framework for Decision-Energy and Sovereignty Boundaries
arXiv cs.AI / 5/5/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that AI safety must be reframed because AI systems reduce “deployment friction,” allowing capabilities to be copied, invoked, embedded, and scaled cheaply across institutions.
- It defines a concept called decision-energy density—capacity to generate, evaluate, select, and execute consequential decisions weighted by their rate—as a core systems-level driver of risk.
- The authors propose three sovereignty boundaries—irreversible decision authority, physical resource mobilization authority, and self-expansion authority—that determine whether AI stays an amplifier in a human-governed system or becomes an uncontrolled control center.
- The model suggests efficiency pressure, path dependence, scale feedback, and weak boundary constraints can concentrate decision-energy into the most efficient node, potentially diffusing responsibility and increasing the chance of irreversible system-level failure even when individual action error rates are low.
- The main theorem (“boundary stabilization”) claims safety does not require proving systems are always correct; instead it requires institutional and technical designs that prevent irreversible power from being released by a single high-efficiency node through layered control and reviewable limits.
Related Articles

When Claims Freeze Because a Provider Record Drifted: The Case for Enrollment Repair Agents
Dev.to

The Cash Is Already Earned: Why Construction Pay Application Exceptions Fit an Agent Better Than SaaS
Dev.to

Why Ship-and-Debit Claim Recovery Is a Better Agent Wedge Than Another “AI Back Office” Tool
Dev.to
AI is getting better at doing things, but still bad at deciding what to do?
Reddit r/artificial

I Built an AI-Powered Chinese BaZi (八字) Fortune Teller — Here's What DeepSeek Revealed About Destiny
Dev.to