Understanding Teacher Revisions of Large Language Model-Generated Feedback
arXiv cs.CL / 3/31/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper studies how 117 teachers revise large-language-model (LLM) formative feedback, analyzing 1,349 paired instances of AI-generated feedback and teacher-edited explanations.
- Teachers accept AI feedback unchanged about 80% of the time, while edits are typically lengthier in draft form and then later shortened by teachers.
- Revision behavior is highly uneven across teachers: roughly half never edit, and only about 10% edit more than two-thirds of instances.
- A model using only the AI feedback text can predict whether teachers will revise with fair accuracy (AUC = 0.75), suggesting revision signals are detectable from the original text.
- When teachers do revise, they often simplify the AI feedback, shifting it from high-information explanations toward more concise, corrective feedback that better matches teacher priorities.



