Diversity in Large Language Models under Supervised Fine-Tuning
arXiv cs.LG / 5/4/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that while supervised fine-tuning (SFT) is crucial for aligning large language models (LLMs) to user intent, it tends to reduce generative diversity, and notes that this effect has not been rigorously tested enough.
- The authors identify two key causes of diversity decline: under-representing low-frequency patterns in the fine-tuning data and forgetting of knowledge learned during pretraining.
- Based on a theoretical analysis, they propose Tempered Focal (TOFU) loss, an SFT training objective designed to address both issues together.
- Large-scale experiments across multiple models and benchmarks show that SFT narrows generation breadth, but TOFU restores or improves diversity while maintaining high response quality.
- Overall, the work provides an empirical and principled method for performing SFT without sacrificing output diversity as much.
Related Articles

When Claims Freeze Because a Provider Record Drifted: The Case for Enrollment Repair Agents
Dev.to

The Cash Is Already Earned: Why Construction Pay Application Exceptions Fit an Agent Better Than SaaS
Dev.to

Why Ship-and-Debit Claim Recovery Is a Better Agent Wedge Than Another “AI Back Office” Tool
Dev.to
AI is getting better at doing things, but still bad at deciding what to do?
Reddit r/artificial

I Built an AI-Powered Chinese BaZi (八字) Fortune Teller — Here's What DeepSeek Revealed About Destiny
Dev.to