AI Navigate

A Quantitative Characterization of Forgetting in Post-Training

arXiv cs.LG / 3/13/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a two-mode mixture abstraction (representing old and new tasks) to theoretically characterize forgetting in continual post-training of generative models, defining two forms: mass forgetting and old-component drift.
  • In the equal-covariance Gaussian setting, forward-KL training on the new distribution drives the old mixture weight to zero (mass forgetting), whereas reverse-KL objectives converge to the target and cause drift only via overlap-gated misassignment, controlled by the Bhattacharyya coefficient with exponential decay as mode separation grows.
  • The authors show how replay interacts with these objectives: for forward-KL, replay must modify the training distribution to change the population optimum; for reverse-KL, replay leaves the objective unchanged but prevents finite-batch old-mode starvation through bounded importance weighting.
  • They analyze three post-training methods (SDFT, TTT-Discover, OAPL) and derive explicit conditions under which each retains old mass or exhibits overlap-controlled drift.
  • Overall, forgetting can be precisely quantified based on the interaction between divergence direction, geometric behavioral overlap, sampling regime, and the visibility of past behavior during training.

Abstract

Continual post-training of generative models is widely used, yet a principled understanding of when and why forgetting occurs remains limited. We develop theoretical results under a two-mode mixture abstraction (representing old and new tasks), proposed by Chen et al. (2025) (arXiv:2510.18874), and formalize forgetting in two forms: (i) mass forgetting, where the old mixture weight collapses to zero, and (ii) old-component drift, where an already-correct old component shifts during training. For equal-covariance Gaussian modes, we prove that forward-KL objectives trained on data from the new distribution drive the old weight to zero, while reverse-KL objectives converge to the true target (thereby avoiding mass forgetting) and perturb the old mean only through overlap-gated misassignment probabilities controlled by the Bhattacharyya coefficient, yielding drift that decays exponentially with mode separation and a locally well-conditioned geometry with exponential convergence. We further quantify how replay interacts with these objectives. For forward-KL, replay must modify the training distribution to change the population optimum; for reverse-KL, replay leaves the population objective unchanged but prevents finite-batch old-mode starvation through bounded importance weighting. Finally, we analyze three recently proposed near-on-policy post-training methods, SDFT (arxiv:2601.19897), TTT-Discover (arxiv:2601.16175), and OAPL (arxiv:2602.19362), via the same lens and derive explicit conditions under which each retains old mass and exhibits overlap-controlled drift. Overall, our results show that forgetting can by precisely quantified based on the interaction between divergence direction, geometric behavioral overlap, sampling regime, and the visibility of past behavior during training.