EGAD: Entropy-Guided Adaptive Distillation for Token-Level Knowledge Transfer

arXiv cs.CL / 5/5/2026

📰 NewsModels & Research

Key Points

  • The paper proposes EGAD, an entropy-guided adaptive knowledge distillation method to improve token-level knowledge transfer from a large LLM teacher to a smaller student model.
  • EGAD addresses a key weakness in prior distillation approaches by treating tokens differently according to their contribution, using the teacher’s output entropy to drive training.
  • It introduces a token-level curriculum that shifts attention from low-entropy tokens to high-entropy tokens during training, plus an entropy-based adjustment of the distillation temperature to reflect the teacher’s confidence.
  • The method uses a dual-branch architecture that performs logits-only distillation for easier tokens while applying deeper feature-based distillation for harder tokens, improving efficiency and learning effectiveness.
  • The authors report extensive experiments showing that EGAD is both sound and effective compared with existing distillation strategies.

Abstract

Large language models (LLMs) have achieved remarkable performance across diverse domains, yet their enormous computational and memory requirements hinder deployment in resource-constrained environments. Knowledge distillation offers a promising solution by transferring knowledge from a large teacher model to a smaller student model. However, existing distillation methods typically treat all tokens equally, ignoring the fact that different tokens contribute unequally to model decisions. This can lead to inefficient knowledge transfer and reduced learning effectiveness. To address this limitation, we propose an entropy-based adaptive distillation strategy that dynamically adjusts the training process at the token level. Our method leverages the teacher's output entropy to guide three aspects of distillation. Specifically, we introduce a token-level curriculum by dynamically shifting focus from low- to high-entropy tokens during training. We further adjust the distillation temperature based on token entropy to better capture teacher confidence patterns. Moreover, we employ a dual-branch architecture for efficient logits-only distillation on easy tokens and deeper feature-based distillation on difficult tokens. Extensive experiments validate the soundness and effectiveness of our method.