Machine Behavior in Relational Moral Dilemmas: Moral Rightness, Predicted Human Behavior, and Model Decisions

arXiv cs.CL / 4/24/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies whether large language models (LLMs) capture social-context effects in moral dilemmas, focusing on the Whistleblower's Dilemma by varying crime severity and relational closeness.
  • It compares three viewpoints—moral rightness (prescriptive norms), predicted human behavior (descriptive expectations), and the model’s own autonomous decisions—to see how each responds to changes in relationship closeness.
  • Results show a strong cross-perspective divergence: moral rightness judgments stay fairness-oriented, while predicted human behavior shifts toward loyalty as relationships become closer.
  • The model’s decisions align with moral rightness rather than with its own predictions about human behavior, indicating reliance on rigid prescriptive rules instead of social nuance from its internal world model.
  • The authors warn that this mismatch could create misalignment risks when such systems are deployed as decision-support tools in real-world social settings.

Abstract

Human moral judgment is context-dependent and modulated by interpersonal relationships. As large language models (LLMs) increasingly function as decision-support systems, determining whether they encode these social nuances is critical. We characterize machine behavior using the Whistleblower's Dilemma by varying two experimental dimensions: crime severity and relational closeness. Our study evaluates three distinct perspectives: (1) moral rightness (prescriptive norms), (2) predicted human behavior (descriptive social expectations), and (3) autonomous model decision-making. By analyzing the reasoning processes, we identify a clear cross-perspective divergence: while moral rightness remains consistently fairness-oriented, predicted human behavior shifts significantly toward loyalty as relational closeness increases. Crucially, model decisions align with moral rightness judgments rather than their own behavioral predictions. This inconsistency suggests that LLM decision-making prioritizes rigid, prescriptive rules over the social sensitivity present in their internal world-modeling, which poses a gap that may lead to significant misalignments in real-world deployments.