Improving Reasoning Capabilities in Small Models through Mixture-of-Layers Distillation with Stepwise Attention on Key Information

arXiv cs.CL / 4/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses the high compute cost of large language models by focusing on distilling reasoning capabilities into smaller models using chain-of-thought (CoT) distillation.
  • It argues that existing CoT distillation approaches largely ignore how a teacher model dynamically shifts attention toward critical information during reasoning.
  • The authors introduce a new CoT distillation framework that transfers the teacher’s stepwise attention on key information to guide the student’s progressive focus.
  • They add a “Mixture of Layers” module to dynamically align different layer representations between teacher and student models.
  • Experiments show consistent improvements on multiple mathematical and commonsense reasoning datasets, and the work claims novelty in leveraging stepwise attention within CoT distillation for small-model reasoning.

Abstract

The significant computational demands of large language models have increased interest in distilling reasoning abilities into smaller models via Chain-of-Thought (CoT) distillation. Current CoT distillation methods mainly focus on transferring teacher-generated rationales for complex reasoning to student models. However, they do not adequately explore teachers' dynamic attention toward critical information during reasoning. We find that language models exhibit progressive attention shifts towards key information during reasoning, which implies essential clues for drawing conclusions. Building on this observation and analysis, we introduce a novel CoT distillation framework that transfers the teacher's stepwise attention on key information to the student model. This establishes structured guidance for the student's progressive concentration on key information during reasoning. More importantly, we develop a Mixture of Layers module enabling dynamic alignment that adapts to different layers between the teacher and student. Our method achieves consistent performance improvements across multiple mathematical and commonsense reasoning datasets. To our knowledge, it is the first method to leverage stepwise attention within CoT distillation to improve small model reasoning.

Improving Reasoning Capabilities in Small Models through Mixture-of-Layers Distillation with Stepwise Attention on Key Information | AI Navigate