Alignment as Institutional Design: From Behavioral Correction to Transaction Structure in Intelligent Systems

arXiv cs.AI / 4/16/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper critiques prevailing AI alignment methods like RLHF as “behavioral correction,” arguing they scale poorly because they resemble an economy that lacks property rights and thus requires continual policing.
  • It proposes a shift to “alignment as institutional design,” where the internal transaction structure of an intelligent system (e.g., module boundaries, competition topology, and cost-feedback loops) is specified so aligned behavior becomes the lowest-cost strategy.
  • Using concepts from institutional economics, the author frames alignment as a political-economy problem rather than a pure behavioral control problem, emphasizing that institutions cannot remove self-interest or guarantee optimality.
  • The work identifies three irreducible human-intervention levels—structural, parametric, and monitorial—and concludes that the objective should be institutional robustness via dynamic, self-correcting processes under oversight.
  • The paper connects its framework to companion research on “Wuxing” resource-competition mechanisms, positioning institutional design as the normative foundation for that approach.

Abstract

Current AI alignment paradigms rely on behavioral correction: external supervisors (e.g., RLHF) observe outputs, judge against preferences, and adjust parameters. This paper argues that behavioral correction is structurally analogous to an economy without property rights, where order requires perpetual policing and does not scale. Drawing on institutional economics (Coase, Alchian, Cheung), capability mutual exclusivity, and competitive cost discovery, we propose alignment as institutional design: the designer specifies internal transaction structures (module boundaries, competition topologies, cost-feedback loops) such that aligned behavior emerges as the lowest-cost strategy for each component. We identify three irreducible levels of human intervention (structural, parametric, monitorial) and show that this framework transforms alignment from a behavioral control problem into a political-economy problem. No institution eliminates self-interest or guarantees optimality; the best design makes misalignment costly, detectable, and correctable. We conclude that the proper goal is institutional robustness-a dynamic, self-correcting process under human oversight, not perfection. This work provides the normative foundation for the Wuxing resource-competition mechanisms in companion papers. Keywords: AI alignment, institutional design, transaction costs, property rights, resource competition, behavioral correction, RLHF, cost truthfulness, modular architecture, correctable alignment