Where does output diversity collapse in post-training?
arXiv cs.CL / 4/20/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- Post-trained language models show reduced output diversity compared with base models, which can weaken inference-time scaling methods that depend on varied samples.
- The study traces output diversity across three post-training lineages (Think/chain-of-thought distillation, Instruct/multi-source data, and RL-Zero) using Olmo 3-style models on 15 tasks and multiple text-diversity metrics.
- The point at which diversity collapses correlates strongly with training data composition: Think loses semantic diversity mainly during supervised fine-tuning, while DPO has a larger effect in Instruct than in Think.
- Disabling chain-of-thought reasoning at inference for Think models reduces accuracy on difficult tasks but does not materially change answer-level diversity, implying the diversity collapse is encoded in model weights from training data rather than caused by generation format.
- By decomposing diversity loss into quality-control (filtering incorrect outputs) and a residual narrowing among correct outputs, the authors find task-dependent behavior and show Think models keep more correct-answer diversity than Instruct overall.
Related Articles

From Theory to Reality: Why Most AI Agent Projects Fail (And How Mine Did Too)
Dev.to

GPT-5.4-Cyber: OpenAI's Game-Changer for AI Security and Defensive AI
Dev.to

Building Digital Souls: The Brutal Reality of Creating AI That Understands You Like Nobody Else
Dev.to
Local LLM Beginner’s Guide (Mac - Apple Silicon)
Reddit r/artificial

Is Your Skill Actually Good? Systematically Validating Agent Skills with Evals
Dev.to