Modeling the human lexicon under temperature variations: linguistic factors, diversity and typicality in LLM word associations
arXiv cs.CL / 3/20/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The study compares human and LLM-generated word associations using the SWOW dataset and three LLMs (Mistral-7B, Llama-3.1-8B, Qwen-2.5-32B) across multiple temperature settings.
- It examines how lexical factors such as word frequency and concreteness influence cue-response pairs in both humans and models.
- Results show all models mirror human trends for frequency and concreteness but differ in response variability and typicality, with larger models emitting highly typical but less variable responses.
- Temperature settings modulate this trade-off by increasing variability while reducing typicality, highlighting how sampling temperature shapes lexical representations.
- The work underscores the importance of considering model size and temperature when probing LLM lexical representations and comparing to human data.
Related Articles
Speaking of VoxtralResearchVoxtral TTS: A frontier, open-weights text-to-speech model that’s fast, instantly adaptable, and produces lifelike speech for voice agents.
Mistral AI Blog
Anyone who has any common sense knows that AI agents in marketing just don’t exist.
Dev.to
How to Use MiMo V2 API for Free in 2026: Complete Guide
Dev.to
The Agent Memory Problem Nobody Solves: A Practical Architecture for Persistent Context
Dev.to
From Chaos to Compliance: AI Automation for the Mobile Kitchen
Dev.to