Assessing Cognitive Biases in LLMs for Judicial Decision Support: Virtuous Victim and Halo Effects
arXiv cs.AI / 3/12/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The study investigates whether large language models exhibit human-like cognitive biases that could affect judicial sentencing decisions, focusing on virtuous victim effects and prestige-based halo effects.
- It uses vignettes modified to avoid training-data recall and evaluates five representative LLMs across multiple trials to isolate each manipulation.
- The findings show a larger virtuous victim effect, no statistically significant penalty for adjacent-consent, and a halo effect that is slightly reduced compared to humans, with credential-based prestige showing the largest reduction.
- Despite cross-model variation, the research suggests only modest improvements over human benchmarks and notes current judicial usage is restricted, underscoring the need for caution and bias mitigation.
Related Articles

Interactive Web Visualization of GPT-2
Reddit r/artificial
Stop Treating AI Interview Fraud Like a Proctoring Problem
Dev.to
[R] Causal self-attention as a probabilistic model over embeddings
Reddit r/MachineLearning
The 5 software development trends that actually matter in 2026 (and what they mean for your startup)
Dev.to
InVideo AI Review: Fast Finished
Dev.to