AI Navigate

Marked Pedagogies: Examining Linguistic Biases in Personalized Automated Writing Feedback

arXiv cs.CL / 3/16/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The study evaluates feedback produced by four LLMs (GPT-4o, GPT-3.5-turbo, Llama-3.3 70B, Llama-3.1 8B) on 600 eighth-grade persuasive essays from the PERSUADE dataset, embedding attributes like gender, race/ethnicity, learning needs, achievement, and motivation in prompts.
  • Using the Marked Words framework, the authors analyze lexical shifts and find systematic stereotype-aligned biases in feedback conditioned on assumed student attributes, even when essay content is identical.
  • The results show biases such as overpraise (positive bias) and withholding substantive critique for students identified by race, language, or disability, potentially reinforcing perceived ability gaps.
  • Beyond content emphasis, the models also shape how writing is judged and how students are addressed, reflecting an instructional orientation the authors term "Marked Pedagogies."
  • The authors call for transparency and accountability in automated feedback tools to mitigate these biases and ensure fair, constructive feedback.

Abstract

Effective personalized feedback is critical to students' literacy development. Though LLM-powered tools now promise to automate such feedback at scale, LLMs are not language-neutral: they privilege standard academic English and reproduce social stereotypes, raising concerns about how "personalization" shapes the feedback students receive. We examine how four widely used LLMs (GPT-4o, GPT-3.5-turbo, Llama-3.3 70B, Llama-3.1 8B) adapt written feedback in response to student attributes. Using 600 eighth-grade persuasive essays from the PERSUADE dataset, we generated feedback under prompt conditions embedding gender, race/ethnicity, learning needs, achievement, and motivation. We analyze lexical shifts across model outputs by adapting the Marked Words framework. Our results reveal systematic, stereotype-aligned shifts in feedback conditioned on presumed student attributes--even when essay content was identical. Feedback for students marked by race, language, or disability often exhibited positive feedback bias and feedback withholding bias--overuse of praise, less substantive critique, and assumptions of limited ability. Across attributes, models tailored not only what content was emphasized but also how writing was judged and how students were addressed. We term these instructional orientations Marked Pedagogies and highlight the need for transparency and accountability in automated feedback tools.