SocioEval: A Template-Based Framework for Evaluating Socioeconomic Status Bias in Foundation Models

arXiv cs.CL / 4/6/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • SocioEval is a template-based evaluation framework designed specifically to measure socioeconomic status (SES) bias in foundation models using decision-making tasks.
  • The framework organizes assessments into 8 themes and 18 topics, producing 240 prompts across 6 class-pair combinations for systematic auditing.
  • Using a three-stage annotation protocol, the authors evaluated 13 frontier LLMs on 3,120 responses and found large variation in bias rates, ranging from 0.42% to 33.75%.
  • The study reports that bias differs by decision type, with lifestyle-related judgments showing about 10× higher bias than education-related decisions.
  • While deployment safeguards reduce explicit discrimination, the results suggest they can be brittle against domain-specific SES stereotypes, and SocioEval is positioned as scalable and extensible for future audits.

Abstract

As Large Language Models (LLMs) increasingly power decision-making systems across critical domains, understanding and mitigating their biases becomes essential for responsible AI deployment. Although bias assessment frameworks have proliferated for attributes such as race and gender, socioeconomic status bias remains significantly underexplored despite its widespread implications in the real world. We introduce SocioEval, a template-based framework for systematically evaluating socioeconomic bias in foundation models through decision-making tasks. Our hierarchical framework encompasses 8 themes and 18 topics, generating 240 prompts across 6 class-pair combinations. We evaluated 13 frontier LLMs on 3,120 responses using a rigorous three-stage annotation protocol, revealing substantial variation in bias rates (0.42\%-33.75\%). Our findings demonstrate that bias manifests differently across themes lifestyle judgments show 10\times higher bias than education-related decisions and that deployment safeguards effectively prevent explicit discrimination but show brittleness to domain-specific stereotypes. SocioEval provides a scalable, extensible foundation for auditing class-based bias in language models.