Language Models Struggle to Use Representations Learned In-Context
arXiv cs.CL / 5/4/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper studies whether large language models can take representations learned from in-context examples and reliably use them for downstream tasks after deployment.
- Experiments show that open-weights LLMs struggle to deploy in-context-defined representations with novel semantics, even when those semantics appear to be captured in latent space.
- The authors evaluate a new benchmark called “adaptive world modeling” and find that closed-source state-of-the-art reasoning models also fail to leverage novel in-context patterns consistently.
- Overall, the work suggests that current LLMs can form in-context representations but lack the ability to flexibly apply them, motivating new methods to improve representation use and transfer.
- The findings highlight a gap between in-context representation learning and the broader goal of adapting behavior to radically new deployment contexts.
Related Articles
AnnouncementsBuilding a new enterprise AI services company with Blackstone, Hellman & Friedman, and Goldman Sachs
Anthropic News

Dara Khosrowshahi on replacing Uber drivers — and himself — with AI
The Verge

CLMA Frame Test
Dev.to

You Are Right — You Don't Need CLAUDE.md
Dev.to

Governance and Liability in AI Agents: What I Built Trying to Answer Those Questions
Dev.to