Sigmoid Head for Quality Estimation under Language Ambiguity
arXiv cs.CL / 3/30/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- This paper argues that standard language-model probabilities (from the softmax head) are unreliable for quality estimation because ambiguous language can yield multiple valid outputs while softmax forces probability mass into a single best option.
- It identifies two core causes: softmax’s inability to assign high probabilities to multiple correct options simultaneously, and training targets being one-hot references that implicitly assume only one correct token.
- The authors introduce a trainable “Sigmoid Head,” an additional unembedding head with sigmoid activation, designed to produce a better quality signal under ambiguity and to remain computationally efficient.
- To address training-target single-reference limitations, they use a negative-sampling heuristic that avoids likely alternative-correct tokens during Sigmoid Head training.
- The paper claims the Sigmoid Head provides a notably improved quality estimate signal versus the original softmax head and is more robust in out-of-domain settings because it does not require human-annotated quality data.
Related Articles

Black Hat Asia
AI Business

Freedom and Constraints of Autonomous Agents — Self-Modification, Trust Boundaries, and Emergent Gameplay
Dev.to
Von Hammerstein’s Ghost: What a Prussian General’s Officer Typology Can Teach Us About AI Misalignment
Reddit r/artificial

Stop Tweaking Prompts: Build a Feedback Loop Instead
Dev.to
Privacy-Preserving Active Learning for autonomous urban air mobility routing under real-time policy constraints
Dev.to