Application-Driven Pedagogical Knowledge Optimization of Open-Source LLMs via Reinforcement Learning and Supervised Fine-Tuning

arXiv cs.CL / 2026/4/9

📰 ニュースSignals & Early TrendsIdeas & Deep AnalysisModels & Research

要点

  • The paper proposes a multi-stage training approach that combines reinforcement learning (RL) and supervised fine-tuning (SFT) to improve LLMs’ pedagogical knowledge for education-focused tasks.
  • The RL stage uses techniques such as progressive difficulty training, emphasis on challenging examples, and extended reasoning rollouts, followed by an SFT stage that distills higher-quality data from the RL-trained model using difficulty-weighted sampling.
  • An optional second RL round is described, creating an extensible pipeline for further pedagogical optimization.
  • Using EduQwen 32B-RL1, EduQwen 32B-SFT, and EduQwen 32B-SFT-RL2 built on a dense Qwen3-32B backbone, the authors report new state-of-the-art results on pedagogical benchmarks (including the interactive Pedagogy Benchmark Leaderboard) and performance that surpasses larger proprietary systems like Gemini-3 Pro.
  • The work argues that domain-specialized optimization can turn mid-sized, open-source LLMs into effective educational domain experts while maintaining transparency, customizability, and cost-efficiency.

Abstract

We present an innovative multi-stage optimization strategy combining reinforcement learning (RL) and supervised fine-tuning (SFT) to enhance the pedagogical knowledge of large language models (LLMs), as illustrated by EduQwen 32B-RL1, EduQwen 32B-SFT, and an optional third-stage model EduQwen 32B-SFT-RL2: (1) RL optimization that implements progressive difficulty training, focuses on challenging examples, and employs extended reasoning rollouts; (2) a subsequent SFT phase that leverages the RL-trained model to synthesize high-quality training data with difficulty-weighted sampling; and (3) an optional second round of RL optimization. EduQwen 32B-RL1, EduQwen 32B-SFT, and EduQwen 32B-SFT-RL2 are an application-driven family of open-source pedagogical LLMs built on a dense Qwen3-32B backbone. These models remarkably achieve high enough accuracy on the Cross-Domain Pedagogical Knowledge (CDPK) Benchmark to establish new state-of-the-art (SOTA) results across the interactive Pedagogy Benchmark Leaderboard and surpass significantly larger proprietary systems such as the previous benchmark leader Gemini-3 Pro. These dense 32-billion-parameter models demonstrate that domain-specialized optimization can transform mid-sized open-source LLMs into true pedagogical domain experts that outperform much larger general-purpose systems, while preserving the transparency, customizability, and cost-efficiency required for responsible educational AI deployment.