SRA: Span Representation Alignment for Large Language Model Distillation

arXiv cs.CL / 5/5/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces SRA, a new framework for Cross-Tokenizer Knowledge Distillation (CTKD) that avoids brittle token-level alignment by aligning at the span level.
  • SRA treats each span as a cluster of particles and represents it using an attention-weighted “center of mass” to capture semantic information in a tokenizer-agnostic way.
  • It prioritizes salient spans via attention-derived weighting and adds a geometric regularizer to maintain the structure of the representation space.
  • The method also includes aligned span logit distillation to improve how knowledge transfers between teacher and student models.
  • In cross-architecture distillation experiments, SRA reportedly delivers consistent and significant improvements over existing CTKD baselines, supporting the authors’ physically grounded modeling approach.

Abstract

Cross-Tokenizer Knowledge Distillation (CTKD) enables knowledge transfer between a large language model and a smaller student, even when they employ different tokenizers. While existing approaches mainly focus on token-level alignment strategies, which are often brittle and sensitive to discrepancies between tokenizers, we argue that the method of aggregating tokens into more robust representations before distillation is of equal importance. In this paper, we introduce \textbf{SRA} (\textbf{S}pan \textbf{R}epresentation \textbf{A}lignment for Large Language Model Distillation), a novel framework that reframes CTKD through the physical lens of Multi-Particle Dynamical Systems. SRA shifts the fundamental unit of alignment from tokens to robust, tokenizer-agnostic spans. We model each span as a cluster of particles and represent its state by its Center of Mass (CoM) - an attention-weighted average that captures rich semantic information. We leverage the concept of span centers of mass with attention-derived weighting to prioritize the most salient spans. In addition, we employ a geometric regularizer to preserve the structural integrity of the representation space and introduce aligned span logit distillation to enhance knowledge transfer across models. In challenging cross-architecture distillation experiments, SRA consistently and significantly outperforms state-of-the-art CTKD baselines, validating our physically-grounded approach.