Sociodemographic Biases in Educational Counselling by Large Language Models

arXiv cs.AI / 4/30/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The study tests six large language models used for educational counselling by having them answer questions about 900 student vignettes across multiple sociodemographic attributes.
  • Results show that all evaluated models exhibit measurable sociodemographic biases, with bias patterns that both resemble known human biases and also differ in important ways.
  • The size of bias depends heavily on how precisely students are described: vague or minimal descriptions can amplify disparities nearly threefold, while concrete and individualized details substantially reduce them.
  • Bias profiles vary widely across different models, indicating that fairness risks are model-dependent and not uniform.
  • The paper argues that more context-rich, personalized student representations can help promote fairness and equity in AI-assisted educational decision-making.

Abstract

As Large Language Models (LLMs) are increasingly integrated into educational settings, understanding their potential biases is critical. This study examines sociodemographic biases in LLM-based educational counselling. We evaluate responses from six LLMs answering questions about 900 vignettes describing students in diverse circumstances. Each vignette is systematically tested across 14 sociodemographic identifiers - spanning race and gender, socioeconomic status, and immigrant background - along with a control condition, yielding 243,000 model responses. Our findings indicate that (1) all models exhibit measurable biases, (2) bias patterns partially align with documented human biases but diverge in notable ways, (3) the magnitude of these biases is strongly influenced by the precision of the student descriptions, where vague or minimal information amplifies disparities nearly threefold, while concrete, individualised metrics substantially reduce them, and (4) bias profiles vary substantially across models. These results demonstrate the importance of context-rich and personalised educational representations, suggesting that AI-driven educational decisions benefit from detailed student-specific information to promote fairness and equity.