Human-in-the-Loop Control of Objective Drift in LLM-Assisted Computer Science Education

arXiv cs.AI / 4/2/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that LLM-assisted programming in computer science education can suffer from “objective drift,” where outputs remain plausible but no longer match task specifications.
  • It reframes human-in-the-loop (HITL) as a durable, teachable control problem (using systems engineering and control-theoretic ideas) rather than a temporary step toward full AI autonomy.
  • The proposed undergraduate CS lab curriculum explicitly separates planning from execution and trains students to set acceptance criteria and architectural constraints before code generation.
  • It also introduces deliberate, concept-aligned drift in some labs to help students diagnose and recover from specification violations.
  • A three-arm pilot study (unstructured AI use vs. structured planning vs. structured planning with injected drift) includes a sensitivity power analysis to estimate detectable effect sizes under realistic class constraints.

Abstract

Large language models (LLMs) are increasingly embedded in computer science education through AI-assisted programming tools, yet such workflows often exhibit objective drift, in which locally plausible outputs diverge from stated task specifications. Existing instructional responses frequently emphasize tool-specific prompting practices, limiting durability as AI platforms evolve. This paper adopts a human-centered stance, treating human-in-the-loop (HITL) control as a stable educational problem rather than a transitional step toward AI autonomy. Drawing on systems engineering and control-theoretic concepts, we frame objectives and world models as operational artifacts that students configure to stabilize AI-assisted work. We propose a pilot undergraduate CS laboratory curriculum that explicitly separates planning from execution and trains students to specify acceptance criteria and architectural constraints prior to code generation. In selected labs, the curriculum also introduces deliberate, concept-aligned drift to support diagnosis and recovery from specification violations. We report a sensitivity power analysis for a three-arm pilot design comparing unstructured AI use, structured planning, and structured planning with injected drift, establishing detectable effect sizes under realistic section-level constraints. The contribution is a theory-driven, methodologically explicit foundation for HITL pedagogy that renders control competencies teachable across evolving AI tools.