Social Meaning in Large Language Models: Structure, Magnitude, and Pragmatic Prompting
arXiv cs.AI / 4/6/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper tests whether large language models capture human social meaning both qualitatively and quantitatively, using new calibration-focused metrics (ESR and CDS) to separate structural fidelity from magnitude calibration.
- Across a case study on numerical (im)precision, frontier LLMs reproduce the qualitative structure of human social inferences but vary widely in how strongly they calibrate the magnitude of those inferences.
- Prompting grounded in pragmatic theory—specifically encouraging reasoning about the speaker’s knowledge state and communicative motives—reduces magnitude deviation more reliably than prompting that focuses on alternative-awareness.
- Combining both pragmatic components improves multiple calibration-sensitive metrics across all evaluated models, though fine-grained magnitude calibration remains only partially resolved.
- Overall, the results suggest LLMs model the inferential structure of pragmatic/social reasoning but still distort inferential strength, and pragmatic-theory prompting helps in a limited, incomplete way.
Related Articles
How Bash Command Safety Analysis Works in AI Systems
Dev.to
How to Get Better Output from AI Tools (Without Burning Time and Tokens)
Dev.to
How I Added LangChain4j Without Letting It Take Over My Spring Boot App
Dev.to
The Future of Artificial Intelligence in Everyday Life
Dev.to
Teaching Your AI to Read: Automating Document Triage for Investigators
Dev.to