Do LLMs Capture Embodied Cognition and Cultural Variation? Cross-Linguistic Evidence from Demonstratives
arXiv cs.CL / 4/29/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes demonstratives (e.g., “this/that” and Chinese “zhe/na”) as a new probe to test whether LLMs can capture grounded, embodied cognition and culture-specific conventions from text.
- Using 6,400 responses from 320 native speakers, the study establishes language-specific human baselines: English speakers distinguish distance reliably but struggle with perspective-taking, while Chinese speakers handle perspective shifts more fluently but accept more distal ambiguity.
- Five state-of-the-art LLMs do not reproduce the human proximal–distal understanding and show no cultural differences, indicating English-centric default reasoning rather than culturally grounded interpretation.
- The authors argue the results inform the egocentric–sociocentric debate and emphasize accounting for individual variation in future model design.
- The contribution includes a new evaluation task and empirical evidence of cross-linguistic asymmetries in how people interpret spatial expressions.
Related Articles

How I Use AI Agents to Maintain a Living Knowledge Base for My Team
Dev.to
IK_LLAMA now supports Qwen3.5 MTP Support :O
Reddit r/LocalLLaMA
OpenAI models, Codex, and Managed Agents come to AWS
Dev.to

Indian Developers: How to Build AI Side Income with $0 Capital in 2026
Dev.to

Vertical SaaS for Startups 2026: Building a Niche AI-First Product
Dev.to