Ideological Bias in LLMs' Economic Causal Reasoning

arXiv cs.CL / 4/24/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The study tests whether large language models (LLMs) show systematic ideological bias when predicting the direction of economic causal effects.
  • It extends the EconCausal benchmark with “ideology-contested” cases where pro-government (intervention-oriented) and pro-market perspectives imply opposite causal signs.
  • Across 20 state-of-the-art LLMs evaluated on 10,490 causality triplets, 1,056 ideology-contested instances were found to be consistently harder than non-contested ones.
  • For 18 of 20 models, prediction accuracy is higher when the empirically verified causal sign matches intervention-oriented expectations, and when models make mistakes, they disproportionately favor intervention-oriented directions.
  • One-shot in-context prompting does not remove this directional skew, suggesting the need for direction-aware evaluation in high-stakes economic and policy use of LLMs.

Abstract

Do large language models (LLMs) exhibit systematic ideological bias when reasoning about economic causal effects? As LLMs are increasingly used in policy analysis and economic reporting, where directionally correct causal judgments are essential, this question has direct practical stakes. We present a systematic evaluation by extending the EconCausal benchmark with ideology-contested cases - instances where intervention-oriented (pro-government) and market-oriented (pro-market) perspectives predict divergent causal signs. From 10,490 causal triplets (treatment-outcome pairs with empirically verified effect directions) derived from top-tier economics and finance journals, we identify 1,056 ideology-contested instances and evaluate 20 state-of-the-art LLMs on their ability to predict empirically supported causal directions. We find that ideology-contested items are consistently harder than non-contested ones, and that across 18 of 20 models, accuracy is systematically higher when the empirically verified causal sign aligns with intervention-oriented expectations than with market-oriented ones. Moreover, when models err, their incorrect predictions disproportionately lean intervention-oriented, and this directional skew is not eliminated by one-shot in-context prompting. These results highlight that LLMs are not only less accurate on ideologically contested economic questions, but systematically less reliable in one ideological direction than the other, underscoring the need for direction-aware evaluation in high-stakes economic and policy settings.