Noise-Response Calibration: A Causal Intervention Protocol for LLM-Judges
arXiv cs.LG / 3/19/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- LLMs are increasingly used as automated judges and synthetic labelers, but their stochasticity and overconfidence complicate deployment when external ground truth is limited.
- The authors propose a practical calibration protocol based on controlled input interventions, asserting that increasing noise severity should lead to a statistically significant deterioration in task performance, evaluated via a slope-based hypothesis test over repeated trials.
- They implement SNR perturbations for tabular data and lexical perturbations for text data, and validate the approach across UCI tabular benchmarks and four text classification datasets, revealing modality-dependent behavior.
- A modality gap is observed: text-based judges degrade predictably while many tabular datasets do not show significant deterioration under noise, and the work provides a reproducible methodology and reporting protocol for robust LLM-judge calibration under distribution shift.
Related Articles
[R] Combining Identity Anchors + Permission Hierarchies achieves 100% refusal in abliterated LLMs — system prompt only, no fine-tuning
Reddit r/MachineLearning
How I Built an AI SDR Agent That Finds Leads and Writes Personalized Cold Emails
Dev.to
Complete Guide: How To Make Money With Ai
Dev.to
I Analyzed My Portfolio with AI and Scored 53/100 — Here's How I Fixed It to 85+
Dev.to
The Demethylation
Dev.to