Why Supervised Fine-Tuning Fails to Learn: A Systematic Study of Incomplete Learning in Large Language Models
arXiv cs.CL / 4/14/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper identifies a persistent failure mode in supervised fine-tuning (SFT) of large language models: even after training convergence, models may not correctly reproduce a subset of their supervised training instances, termed the Incomplete Learning Phenomenon (ILP).
- ILP is shown to be widespread across multiple LLM families, domains, and datasets, and aggregate evaluation metrics can hide these persistent “unlearned” subsets.
- The authors formalize ILP as post-training failure to internalize supervised instances and propose a diagnostic-first framework that classifies unlearned samples into observable, recurrent causes.
- Five key sources of incomplete learning are identified: missing prerequisite knowledge, conflicts with pre-training knowledge, internal inconsistencies in SFT data, left-side forgetting during sequential fine-tuning, and insufficient optimization for rare or complex patterns.
- The study also examines mitigation strategies as causal interventions, using experiments on models including Qwen, LLaMA, and OLMo2 to demonstrate heterogeneous behavior and targeted improvements.
Related Articles

Black Hat Asia
AI Business

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Don't forget, there is more than forgetting: new metrics for Continual Learning
Dev.to

Microsoft MAI-Image-2-Efficient Review 2026: The AI Image Model Built for Production Scale
Dev.to
Bit of a strange question?
Reddit r/artificial