The Crutch or the Ceiling? How Different Generations of LLMs Shape EFL Student Writings

arXiv cs.AI / 4/20/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper investigates how LLM assistance affects secondary EFL students’ writing, focusing on whether newer, smarter models function as scaffolding or become “crutches” that conceal students’ real abilities.
  • By comparing LLM-assisted compositions before and after ChatGPT’s release and using both expert qualitative scoring and quantitative readability/lexical measures, the study finds improvements in scores—especially for lower-proficiency learners.
  • However, greater LLM help is linked to lower human expert ratings, indicating that students may produce fluent-looking text with weaker deep coherence.
  • The authors argue that effective AI-assisted writing pedagogy should shift from optimizing output quality to verifying the learning process, including separating ideational scaffolding from textual production within learners’ Zone of Proximal Development.

Abstract

The rapid evolution of Large Language Models (LLMs) has made them powerful tools for enhancing student writing. This study explores the extent and limitations of LLMs in assisting secondary-level English as a Foreign Language (EFL) students with their writing tasks. While existing studies focus on output quality, our research examines the developmental shift in LLMs and their impact on EFL students, assessing whether smarter models act as true scaffolds or mere compensatory crutches. To achieve this, we analyse student compositions assisted by LLMs before and after ChatGPT's release, using both expert qualitative scoring and quantitative metrics (readability tests, Pearson's correlation coefficient, MTLD, and others). Our results indicate that advanced LLMs boost assessment scores and lexical diversity for lower-proficiency learners, potentially masking their true ability. Crucially, increased LLM assistance correlated negatively with human expert ratings, suggesting surface fluency without deep coherence. To transform AI-assisted practice into genuine learning, pedagogy must shift from focusing on output quality to verifying the learning process. Educators should align AI functions, specifically differentiating ideational scaffolding from textual production, within the learner's Zone of Proximal Development.