Behavioural feasible set: Value alignment constraints on AI decision support
arXiv cs.AI / 2026/3/24
💬 オピニオンSignals & Early TrendsIdeas & Deep AnalysisModels & Research
要点
- The paper argues that when organizations use commercial AI for decision support, they inherit opaque, vendor-embedded value judgments that constrain what recommendations the system can produce.
- It introduces the concept of a “behavioural feasible set,” defining the set of reachable recommendations under vendor-imposed alignment constraints and providing diagnostics for when organizational requirements exceed that flexibility.
- Through scenario-based experiments (binary decision scenarios and multi-stakeholder ranking tasks), the author finds that alignment significantly compresses the feasible set, making recommendations less adjustable under contextual pressure.
- Experiments comparing pre- and post-alignment variants of an open-weight model suggest alignment is the mechanism of increased rigidity, and leading commercial models show similar or stronger effects.
- In multi-stakeholder settings, alignment changes implied stakeholder priorities rather than simply neutralizing them, creating a governance problem that prompting alone cannot fix because vendor choice determines which trade-offs are negotiable.

