Teaching Language Models How to Code Like Learners: Conversational Serialization for Student Simulation
arXiv cs.AI / 4/14/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes a way to train open-weight “programming learner” language models by converting real student temporal log traces into a dialogue-style conversational serialization.
- In the serialized format, alternating turns capture student code submissions and automated assessment/environment feedback (tests, grades, and error traces) to teach models iterative debugging behavior.
- The training pipeline combines supervised fine-tuning with preference optimization to better align the learner model’s responses with authentic student debugging patterns.
- Experiments on Qwen models (4B and 8B) trained with real Python assignment submission data show that including environment feedback improves functional alignment and code similarity versus code-only and prompted LLM baselines.
- The authors release code to support reproducibility and to reduce reliance on proprietary prompting approaches for large-scale tutoring strategy evaluation.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat Asia
AI Business
Microsoft launches MAI-Image-2-Efficient, a cheaper and faster AI image model
VentureBeat

The AI School Bus Camera Company Blanketing America in Tickets
Dev.to
GPT-5.3 and GPT-5.4 on OpenClaw: Setup and Configuration...
Dev.to
GLM-5 on OpenClaw: Setup Guide, Benchmarks, and When to...
Dev.to