A Quantitative Characterization of Forgetting in Post-Training
arXiv cs.LG / 3/13/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces a two-mode mixture abstraction (representing old and new tasks) to theoretically characterize forgetting in continual post-training of generative models, defining two forms: mass forgetting and old-component drift.
- In the equal-covariance Gaussian setting, forward-KL training on the new distribution drives the old mixture weight to zero (mass forgetting), whereas reverse-KL objectives converge to the target and cause drift only via overlap-gated misassignment, controlled by the Bhattacharyya coefficient with exponential decay as mode separation grows.
- The authors show how replay interacts with these objectives: for forward-KL, replay must modify the training distribution to change the population optimum; for reverse-KL, replay leaves the objective unchanged but prevents finite-batch old-mode starvation through bounded importance weighting.
- They analyze three post-training methods (SDFT, TTT-Discover, OAPL) and derive explicit conditions under which each retains old mass or exhibits overlap-controlled drift.
- Overall, forgetting can be precisely quantified based on the interaction between divergence direction, geometric behavioral overlap, sampling regime, and the visibility of past behavior during training.
Related Articles

Attacks On Data Centers, Qwen3.5 In All Sizes, DeepSeek’s Huawei Play, Apple’s Multimodal Tokenizer
The Batch

Your AI generated code is "almost right", and that is actually WORSE than it being "wrong".
Dev.to

Lessons from Academic Plagiarism Tools for SaaS Product Development
Dev.to

**Core Allocation Optimization for Energy‑Efficient Multi‑Core Scheduling in ARINC650 Systems**
Dev.to

KI in der amtlichen Recherche beim DPMA: Was Patentanwälte bei Neuanmeldungen jetzt beachten sollten (Stand: März 2026)
Dev.to