Behavioural feasible set: Value alignment constraints on AI decision support

arXiv cs.AI / 3/24/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that when organizations use commercial AI for decision support, they inherit opaque, vendor-embedded value judgments that constrain what recommendations the system can produce.
  • It introduces the concept of a “behavioural feasible set,” defining the set of reachable recommendations under vendor-imposed alignment constraints and providing diagnostics for when organizational requirements exceed that flexibility.
  • Through scenario-based experiments (binary decision scenarios and multi-stakeholder ranking tasks), the author finds that alignment significantly compresses the feasible set, making recommendations less adjustable under contextual pressure.
  • Experiments comparing pre- and post-alignment variants of an open-weight model suggest alignment is the mechanism of increased rigidity, and leading commercial models show similar or stronger effects.
  • In multi-stakeholder settings, alignment changes implied stakeholder priorities rather than simply neutralizing them, creating a governance problem that prompting alone cannot fix because vendor choice determines which trade-offs are negotiable.

Abstract

When organisations adopt commercial AI systems for decision support, they inherit value judgements embedded by vendors that are neither transparent nor renegotiable. The governance puzzle is not whether AI can support decisions but which recommendations the system can actually produce given how its vendor has configured it. I formalise this as a behavioural feasible set, the range of recommendations reachable under vendor-imposed alignment constraints, and characterise diagnostic thresholds for when organisational requirements exceed the system's flexibility. In scenario-based experiments using binary decision scenarios and multi-stakeholder ranking tasks, I show that alignment materially compresses this set. Comparing pre- and post-alignment variants of an open-weight model isolates the mechanism: alignment makes the system substantially less able to shift its recommendation even under legitimate contextual pressure. Leading commercial models exhibit comparable or greater rigidity. In multi-stakeholder tasks, alignment shifts implied stakeholder priorities rather than neutralising them, meaning organisations adopt embedded value orientations set upstream by the vendor. Organisations thus face a governance problem that better prompting cannot resolve: selecting a vendor partially determines which trade-offs remain negotiable and which stakeholder priorities are structurally embedded.