Self-Distillation as a Performance Recovery Mechanism for LLMs: Counteracting Compression and Catastrophic Forgetting

arXiv cs.LG / 4/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a performance recovery framework for LLMs called Self-Distillation Fine-Tuning (SDFT) to restore capabilities degraded by issues like catastrophic forgetting during SFT, as well as by quantization and pruning.
  • It argues that an LLM’s generative ability is fundamentally tied to the high-dimensional manifold formed by its hidden layers, and that recovery works by realigning this manifold.
  • The authors provide a theoretical explanation connecting recovery to geometric representation ideas, rather than relying only on empirical results.
  • Using Centered Kernel Alignment (CKA), they measure alignment between teacher and student activation trajectories and find performance recovery strongly correlates with manifold alignment.
  • Overall, the study links practical self-distillation methods with geometric representation theory to clarify how self-distillation restores model performance internally.

Abstract

Large Language Models (LLMs) have achieved remarkable success, underpinning diverse AI applications. However, they often suffer from performance degradation due to factors such as catastrophic forgetting during Supervised Fine-Tuning (SFT), quantization, and pruning. In this work, we introduce a performance recovery framework based on Self-Distillation Fine-Tuning (SDFT) that effectively restores model capabilities. Complementing this practical contribution, we provide a rigorous theoretical explanation for the underlying recovery mechanism. We posit that an LLM's generative capability fundamentally relies on the high-dimensional manifold constructed by its hidden layers. To investigate this, we employ Centered Kernel Alignment (CKA) to quantify the alignment between student and teacher activation trajectories, leveraging its invariance to orthogonal transformations and scaling. Our experiments demonstrate a strong correlation between performance recovery and manifold alignment, substantiating the claim that self-distillation effectively aligns the student's high-dimensional manifold with the optimal structure represented by the teacher. This study bridges the gap between practical recovery frameworks and geometric representation theory, offering new insights into the internal mechanisms of self-distillation.