From If-Statements to ML Pipelines: Revisiting Bias in Code-Generation

arXiv cs.CL / 4/24/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that evaluating code-generation bias using only simple if-statements can miss a large portion of real-world bias, because it reflects only a narrow slice of programming behavior.
  • It studies a more realistic scenario—LLM-generated machine learning (ML) pipelines—and shows that bias emerges strongly during feature selection.
  • Across both code-specialized and general-instruction LLMs, sensitive attributes appear in generated pipelines in 87.7% of cases on average, even when irrelevant features are excluded.
  • Compared with if-statement-based conditional evaluations (where sensitive attributes appear in 59.2% of cases), the ML-pipeline setting reveals substantially higher bias rates.
  • The findings remain consistent under different prompt-mitigation strategies and across attribute counts and pipeline difficulty levels, indicating current benchmarks likely understate deployment risk.

Abstract

Prior work evaluates code generation bias primarily through simple conditional statements, which represent only a narrow slice of real-world programming and reveal solely overt, explicitly encoded bias. We demonstrate that this approach dramatically underestimates bias in practice by examining a more realistic task: generating machine learning (ML) pipelines. Testing both code-specialized and general-instruction large language models, we find that generated pipelines exhibit significant bias during feature selection. Sensitive attributes appear in 87.7% of cases on average, despite models demonstrably excluding irrelevant features (e.g., including "race" while dropping "favorite color" for credit scoring). This bias is substantially more prevalent than that captured by conditional statements, where sensitive attributes appear in only 59.2% of cases. These findings are robust across prompt mitigation strategies, varying numbers of attributes, and different pipeline difficulty levels. Our results challenge simple conditionals as valid proxies for bias evaluation and suggest current benchmarks underestimate bias risk in practical deployments.