Hierarchical Apprenticeship Learning from Imperfect Demonstrations with Evolving Rewards

arXiv cs.AI / 4/2/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that apprenticeship learning in e-learning settings should treat imperfect and evolving student demonstrations as structured signals rather than noise to ignore, as long as their relative quality is ranked.
  • It introduces HALIDE (Hierarchical Apprenticeship Learning from Imperfect Demonstrations with Evolving Rewards), which learns from sub-optimal demonstrations by using a hierarchical model to infer higher-level intent and strategy from lower-level actions.
  • HALIDE explicitly captures temporal evolution in student reward functions, helping separate transient mistakes from persistent suboptimal strategies and genuine progress toward learning goals.
  • The authors report that HALIDE more accurately predicts students’ pedagogical decisions than methods that use only optimal trajectories, assume fixed rewards, or treat imperfect demonstrations as unranked.

Abstract

While apprenticeship learning has shown promise for inducing effective pedagogical policies directly from student interactions in e-learning environments, most existing approaches rely on optimal or near-optimal expert demonstrations under a fixed reward. Real-world student interactions, however, are often inherently imperfect and evolving: students explore, make errors, revise strategies, and refine their goals as understanding develops. In this work, we argue that imperfect student demonstrations are not noise to be discarded, but structured signals-provided their relative quality is ranked. We introduce HALIDE, Hierarchical Apprenticeship Learning from Imperfect Demonstrations with Evolving Rewards, which not only leverages sub-optimal student demonstrations, but ranks them within a hierarchical learning framework. HALIDE models student behavior at multiple levels of abstraction, enabling inference of higher-level intent and strategy from suboptimal actions while explicitly capturing the temporal evolution of student reward functions. By integrating demonstration quality into hierarchical reward inference,HALIDE distinguishes transient errors from suboptimal strategies and meaningful progress toward higher-level learning goals. Our results show that HALIDE more accurately predicts student pedagogical decisions than approaches that rely on optimal trajectories, fixed rewards, or unranked imperfect demonstrations.