Metaphors We Compute By: A Computational Audit of Cultural Translation vs. Thinking in LLMs
arXiv cs.CL / 4/7/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that LLMs being able to produce multilingual text does not imply they can perform culture-aware reasoning, especially in creative tasks tied to cultural conceptual frameworks.
- It presents a computational audit using a metaphor generation benchmark across five cultural settings and multiple abstract concepts to test whether LLMs behave as culturally diverse partners or as “translators” anchored in a dominant (not culture-specific) framework.
- The empirical results show stereotyped metaphor patterns for certain cultural settings and evidence of “Western defaultism.”
- The authors conclude that adding a cultural identity to prompts is insufficient to guarantee culturally grounded reasoning, indicating a need for more robust evaluation and mitigation of cultural bias.
Related Articles

Black Hat Asia
AI Business
Research with ChatGPT
Dev.to
Silicon Valley is quietly running on Chinese open source models and almost nobody is talking about it
Reddit r/LocalLLaMA

Why AI Product Quality Is Now an Evaluation Pipeline Problem, Not a Model Problem
Dev.to

The 10 Best AI Tools for SEO and Digital Marketing in 2026
Dev.to