Quantifying and Mitigating Self-Preference Bias of LLM Judges
arXiv cs.LG / 4/28/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- LLM-as-a-Judge evaluation can be distorted by Self-Preference Bias (SPB), where an LLM systematically favors or disfavors outputs it generated during judging.
- Prior SPB measurements are often too expensive and may mix up a model’s generative ability with its evaluative stance, limiting practical large-scale use.
- The paper proposes a fully automated framework that builds equal-quality response pairs to statistically separate true discriminability from bias propensity without needing human gold annotations.
- Experiments on 20 mainstream LLMs suggest that higher capabilities are frequently uncorrelated with (or even negatively correlated with) low SPB, meaning strong models are not automatically fair judges.
- To mitigate SPB, the authors introduce a structured, multi-dimensional evaluation approach based on cognitive load decomposition, reducing SPB by an average of 31.5%.
Related Articles

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Same Agent, Different Risk | How Microsoft 365 Copilot Grounding Changes the Security Model | Rahsi Framework™
Dev.to

Claude Haiku for Low-Cost AI Inference: Patterns from a Horse Racing Prediction System
Dev.to

How We Built an Ambient AI Clinical Documentation Pipeline (and Saved Doctors 8+ Hours a Week)
Dev.to

🦀 PicoClaw Deep Dive — A Field Guide to Building an Ultra-Light AI Agent in Go 🐹
Dev.to