Learning from Supervision with Semantic and Episodic Memory: A Reflective Approach to Agent Adaptation
arXiv cs.CL / 5/4/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes a memory-augmented framework for LLM-based agents to learn target classification functions from labeled examples without updating model parameters.
- It uses semantic memory to convert LLM-generated, label-grounded critiques into reusable task-level guidance, and episodic memory to store instance-level critiques tied to past experiences.
- Experiments across multiple tasks and models show the best self-critique strategy improves accuracy by 8.1 percentage points over a zero-shot baseline and by 4.6 points over a label-only RAG baseline.
- The authors introduce a new metric, “suggestibility,” to explain why performance gains differ significantly by model and domain, identifying when memory augmentation helps or fails.
- They also find that pre-computing critiques reduces inference-time reasoning costs, cutting “thinking” tokens by an average of 31.95% compared with letting the model reason independently.
Related Articles
AnnouncementsBuilding a new enterprise AI services company with Blackstone, Hellman & Friedman, and Goldman Sachs
Anthropic News

Dara Khosrowshahi on replacing Uber drivers — and himself — with AI
The Verge

CLMA Frame Test
Dev.to

You Are Right — You Don't Need CLAUDE.md
Dev.to

Governance and Liability in AI Agents: What I Built Trying to Answer Those Questions
Dev.to