When Can We Trust LLM Graders? Calibrating Confidence for Automated Assessment

arXiv cs.CL / 4/1/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper focuses on when LLM graders can be trusted by predicting the likelihood that their automated grading is correct, enabling selective automation with human review for low-confidence cases.
  • It evaluates three confidence estimation approaches—self-reported confidence, self-consistency voting, and token probability—across seven LLMs (4B–120B) on three educational grading datasets.
  • Self-reported confidence provides the best calibration overall, outperforming self-consistency (lower average ECE) even though self-consistency is more expensive computationally.
  • Larger models generally calibrate better, with GPT-OSS-120B achieving the best average calibration (avg ECE 0.100) and meaningful discrimination (avg AUC 0.668).
  • The authors find confidence scores are strongly top-skewed across methods, producing a “confidence floor” that practitioners must consider when choosing grading thresholds.

Abstract

Large Language Models (LLMs) show promise for automated grading, but their outputs can be unreliable. Rather than improving grading accuracy directly, we address a complementary problem: \textit{predicting when an LLM grader is likely to be correct}. This enables selective automation where high-confidence predictions are processed automatically while uncertain cases are flagged for human review. We compare three confidence estimation methods (self-reported confidence, self-consistency voting, and token probability) across seven LLMs of varying scale (4B to 120B parameters) on three educational datasets: RiceChem (long-answer chemistry), SciEntsBank, and Beetle (short-answer science). Our experiments reveal that self-reported confidence consistently achieves the best calibration across all conditions (avg ECE 0.166 vs 0.229 for self-consistency). Surprisingly, self-consistency remains 38\% worse despite requiring 5\times the inference cost. Larger models exhibit substantially better calibration though gains vary by dataset and method (e.g., a 28\% ECE reduction for self-reported), with GPT-OSS-120B achieving the best calibration (avg ECE 0.100) and strong discrimination (avg AUC 0.668). We also observe that confidence is strongly top-skewed across methods, creating a ``confidence floor'' that practitioners must account for when setting thresholds. These findings suggest that simply asking LLMs to report their confidence provides a practical approach for identifying reliable grading predictions. Code is available \href{https://github.com/sonkar-lab/llm_grading_calibration}{here}.