Modeling the human lexicon under temperature variations: linguistic factors, diversity and typicality in LLM word associations
arXiv cs.CL / 3/20/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The study compares human and LLM-generated word associations using the SWOW dataset and three LLMs (Mistral-7B, Llama-3.1-8B, Qwen-2.5-32B) across multiple temperature settings.
- It examines how lexical factors such as word frequency and concreteness influence cue-response pairs in both humans and models.
- Results show all models mirror human trends for frequency and concreteness but differ in response variability and typicality, with larger models emitting highly typical but less variable responses.
- Temperature settings modulate this trade-off by increasing variability while reducing typicality, highlighting how sampling temperature shapes lexical representations.
- The work underscores the importance of considering model size and temperature when probing LLM lexical representations and comparing to human data.
Related Articles
How AI is Transforming Dynamics 365 Business Central
Dev.to
Algorithmic Gaslighting: A Formal Legal Template to Fight AI Safety Pivots That Cause Psychological Harm
Reddit r/artificial
Do I need different approaches for different types of business information errors?
Dev.to
ShieldCortex: What We Learned Protecting AI Agent Memory
Dev.to
How AI-Powered Revenue Intelligence Transforms B2B Sales Teams
Dev.to