From Domain Understanding to Design Readiness: a playbook for GenAI-supported learning in Software Engineering

arXiv cs.AI / 4/2/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper reports a two-week master’s course milestone where 29 students used a customized GPT-3.5 tutor grounded in a curated knowledge base to learn cryptocurrency-finance basics and Domain-Driven Design (DDD).
  • An evaluation of 60 prompt–answer pairs found responses were highly accurate (98.9% average accuracy, no factual errors; only 2 minor inaccuracies) and relevant (92.2% average), with strong pedagogical value (89.4%) and generally appropriate cognitive load (82.78%).
  • Despite strong content quality, the study found low “supportiveness” (37.78%), indicating that while answers were correct and instructional, the conversational tone and follow-up structure were not as helpful as intended.
  • Students showed large gains in self-efficacy from pre to post for genAI-assisted domain learning and applying DDD, suggesting measurable learning impact.
  • The authors distill 17 concrete teaching practices for GenAI-supported software engineering education, focusing on prompt/configuration and course workflow design (e.g., expected granularity, limiting verbosity, curated guardrail examples, and lightweight quality rubrics).

Abstract

Software engineering courses often require rapid upskilling in supporting knowledge areas such as domain understanding and modeling methods. We report an experience from a two-week milestone in a master's course where 29 students used a customized ChatGPT (GPT-3.5) tutor grounded in a curated course knowledge base to learn cryptocurrency-finance basics and Domain-Driven Design (DDD). We logged all interactions and evaluated a 34.5% random sample of prompt-answer pairs (60/~174) with a five-dimension rubric (accuracy, relevance, pedagogical value, cognitive load, supportiveness), and we collected pre/post self-efficacy. Responses were consistently accurate and relevant in this setting: accuracy averaged 98.9% with no factual errors and only 2/60 minor inaccuracies, and relevance averaged 92.2%. Pedagogical value was high (89.4%) with generally appropriate cognitive load (82.78%), but supportiveness was low (37.78%). Students reported large pre-post self-efficacy gains for genAI-assisted domain learning and DDD application. From these observations we distill seventeen concrete teaching practices spanning prompt/configuration and course/workflow design (e.g., setting expected granularity, constraining verbosity, curating guardrail examples, adding small credit with a simple quality rubric). Within this single-course context, results suggest that genAI-supported learning can complement instruction in domain understanding and modeling tasks, while leaving room to improve tone and follow-up structure.