Interpretable Difficulty-Aware Knowledge Tracing in Tutor-Student Dialogues

arXiv cs.CL / 5/5/2026

📰 NewsModels & Research

Key Points

  • The paper addresses the challenge of knowledge tracing (KT) in tutor–student dialogues by proposing a framework that updates students’ knowledge state at each conversational turn.
  • It improves over prior dialogue-based KT methods by explicitly modeling both question/task difficulty and student ability, rather than relying on opaque latent representations from LLMs.
  • The framework uses Item Response Theory (IRT) to translate LLM outputs into interpretable parameters for student ability and question difficulty, grounding predictions in cognitive learning theory.
  • Experiments on two tutor–student dialogue datasets show the proposed approach outperforms existing KT baselines while producing qualitative, theory-consistent interpretability.
  • Overall, the work advances interpretable, difficulty-aware AI tutoring analytics by combining LLMs with structured cognitive modeling.

Abstract

Recent advances in large language models (LLMs) have led to the development of AI-powered tutoring systems that provide interactive support via dialogue. To enable these tutoring systems to provide personalized support, it is essential to assess student performance at each turn, motivating knowledge tracing (KT) in dialogue settings. However, existing dialogue-based KT approaches often ignore question difficulty modeling and rely on opaque latent representations from LLMs, hindering accurate and interpretable prediction. In this work, we propose an interpretable difficulty-aware conversational KT framework built upon LLMs, which explicitly models students' abilities and the difficulty of tutor-posed tasks at each turn. The framework incorporates the original textual question and the next tutor-posed task to estimate the student's knowledge state and the difficulty of the upcoming turn. Furthermore, it integrates Item Response Theory to map LLM's outputs into student ability and question difficulty parameters, enabling interpretable prediction of student performance grounded in cognitive theories of learning. We evaluate the framework on two tutor-student dialogue datasets. Both quantitative and qualitative results show that our framework outperforms existing KT baselines, meanwhile generating interpretable outputs consistent with cognitive theory.