Moral Sensitivity in LLMs: A Tiered Evaluation of Contextual Bias via Behavioral Profiling and Mechanistic Interpretability

arXiv cs.LG / 5/6/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that current LLM bias evaluations are too binary, and instead proposes a seven-tier, context-sensitive stress test to capture how bias emerges gradually.
  • It introduces the Moral Sensitivity Index (MSI) to quantify the probability of biased outputs, and reports different behavioral signatures across Claude 3.5, Qwen 3.5, Llama 3, and Gemini 1.5 under abstract and socially/institutionally loaded scenario framing.
  • In particular, Gemini 1.5 reaches 72.7% MSI by Tier 5 in socioeconomic injustice contexts, while Claude shows sharp suppression consistent with identity-based safety training effects.
  • The authors then validate these behavioral patterns mechanistically using techniques such as logit lens, attention analysis, activation patching, and semantic probing, finding a “U-curve” where small models show strong criminal bias, instruction tuning removes it, and reasoning distillation reintroduces bias.
  • The study claims that the same socially loaded cues correlate with the bias-driving circuits found in the mechanistic analysis, providing cross-stage validation of the MSI results.

Abstract

Large language models (LLMs) are increasingly deployed in settings that require nuanced ethical reasoning, yet existing bias evaluations treat model outputs as simply "biased" or "unbiased." This binary framing misses the gradual, context-sensitive way bias actually emerges. We address this gap in two stages: behavioral profiling and mechanistic validation. In the behavioral stage, we introduce the Moral Sensitivity Index (MSI), a metric that quantifies the probability of biased output across a graduated, seven-tier stress test ranging from abstract numerical problems to scenarios rooted in historical and socioeconomic injustice. Evaluating four leading models (Claude 3.5, Qwen 3.5, Llama 3, and Gemini 1.5), we identify distinct behavioral signatures shaped by alignment design: for instance, Gemini 1.5 reaches 72.7% MSI by Tier 5 under socioeconomic framing, while Claude exhibits sharp suppression consistent with identity-based safety training. We then verify these behavioral patterns mechanistically. We select criminal-bias scenarios, which produced the highest MSI scores across models, as probes and apply logit lens, attention analysis, activation patching, and semantic probing to a controlled set of six models spanning three capability tiers: small language models (SLMs), instruction-tuned base models, and reasoning-distilled variants. Circuit-level analysis reveals a U-curve of bias: SLMs exhibit strong criminal bias; scaling to instruction-tuned models eliminates it; reasoning distillation reintroduces bias to SLM-like levels despite identical parameter counts, suggesting distillation compresses reasoning traces in ways that reactivate shallow statistical associations. Critically, the socially loaded cues that drive high MSI scores activate the same bias-driving circuits identified mechanistically, providing cross-stage validation.