ADVICE: Answer-Dependent Verbalized Confidence Estimation
arXiv cs.CL / 5/4/2026
💬 OpinionModels & Research
Key Points
- The paper studies why LLMs that verbalize confidence in natural language often become systematically overconfident.
- It identifies “answer-independence”—confidence that does not condition on the model’s own answer—as a key driver of the miscalibration.
- The authors propose ADVICE (Answer-Dependent Verbalized Confidence Estimation), a fine-tuning approach designed to make confidence grounded in the model’s answer.
- Experiments show that ADVICE improves confidence calibration substantially and generalizes to unseen settings without hurting task performance.
- The improvements are attributed to increased answer dependence, offering insight into the mechanisms behind overconfident, more trustworthy confidence verbalization.
Related Articles

ALM on Power Platform: ADO + GitHub, the best of both worlds
Dev.to

Experiment: Does repeated usage influence ChatGPT 5.4 outputs in a RAG-like setup?
Dev.to

When a memorized rule fits your bug too well: a meta-trap of agent workflows
Dev.to
LWiAI Podcast #243 - GPT 5.5, DeepSeek V4, AI safety sabotage
Last Week in AI
Excellent discussion about LLM scaling [D]
Reddit r/MachineLearning