Human-like Working Memory Interference in Large Language Models
arXiv cs.LG / 4/14/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper examines why large language models (LLMs) exhibit working-memory limitations even though transformers can attend to full prior context.
- Experiments show that pretrained LLMs reproduce human-like interference patterns, including performance degradation under higher memory load and biases driven by recency and stimulus statistics.
- A key result is that LLMs rely on interference control over entangled representations of multiple memory items, rather than directly copying the target item from the context.
- The authors provide causal evidence via targeted intervention: suppressing stimulus-content information improves working-memory task performance.
- Across models, stronger working-memory capacity correlates with broader benchmark competence, suggesting working memory as a shared computational constraint linked to general intelligence.
