Diversity in Large Language Models under Supervised Fine-Tuning

arXiv cs.LG / 5/4/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that while supervised fine-tuning (SFT) is crucial for aligning large language models (LLMs) to user intent, it tends to reduce generative diversity, and notes that this effect has not been rigorously tested enough.
  • The authors identify two key causes of diversity decline: under-representing low-frequency patterns in the fine-tuning data and forgetting of knowledge learned during pretraining.
  • Based on a theoretical analysis, they propose Tempered Focal (TOFU) loss, an SFT training objective designed to address both issues together.
  • Large-scale experiments across multiple models and benchmarks show that SFT narrows generation breadth, but TOFU restores or improves diversity while maintaining high response quality.
  • Overall, the work provides an empirical and principled method for performing SFT without sacrificing output diversity as much.

Abstract

Supervised Fine-Tuning (SFT) is essential for aligning Large Language Models (LLMs) with user intent, yet it is believed to suppress generative diversity. Although this reduction is frequently referenced, formal empirical testing of the phenomenon remains limited. The expressiveness of LLMs by itself was addressed by multiple prior methods. Their varying perspectives suggest that deeper analysis could yield further improvements. In this study, we attribute the decline to two primary drivers: the neglect of low-frequency patterns within fine-tuning datasets and the forgetting of preexisting knowledge. Motivated by our theoretical analysis, we develop Tempered Focal (TOFU) loss, a novel objective that addresses both stated challenges simultaneously. Our extensive evaluation confirms at scale that generation breadth narrows after SFT and strengthens the hypothesis explaining this effect. Across multiple models and benchmarks, we demonstrate that TOFU enhances output diversity while preserving high response quality, offering a principled approach to SFT.