Slot Machines: How LLMs Keep Track of Multiple Entities
arXiv cs.CL / 4/24/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper examines how language models represent and maintain bindings between multiple entities and their attributes across token positions within a context.
- It introduces a “multi-slot probing” method to disentangle a token’s residual stream into separate “current-entity” and “prior-entity” components.
- The authors find that the current-entity and prior-entity slots play different functional roles: the current-entity slot supports explicit factual retrieval, while the prior-entity slot better supports relational inferences and conflict detection.
- Open-weight models struggle with syntax that forces two subject-verb-object bindings onto a single token, while newer frontier models handle it, implying improved binding strategies.
- The results highlight a gap between information present in model activations and information the model actually uses, and suggest the current/prior-entity slot structure could support multi-perspective behaviors like sycophancy or deception.
Related Articles

Your MCP server probably has too many tools
Dev.to

MCP Auth That Actually Works: OAuth for Remote Servers
Dev.to

GoDavaii's Day 5: When 22 Indian Languages Redefine 'Hard' in Health AI
Dev.to

Gemma 4 and Qwen 3.6 with q8_0 and q4_0 KV cache: KL divergence results
Reddit r/LocalLLaMA
Corea arresta a hombre por imagen IA falsa del lobo Neukgu: hasta 5 años
Dev.to