Poison Once, Exploit Forever: Environment-Injected Memory Poisoning Attacks on Web Agents
arXiv cs.AI / 4/6/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that LLM-based web agents’ memory—used to personalize across tasks—creates a persistent, cross-session attack surface that can be exploited beyond traditional direct memory tampering assumptions.
- It introduces eTAMP (Environment-injected Trajectory-based Agent Memory Poisoning), showing that an attacker can poison an agent’s stored memory via environmental observation alone (e.g., a manipulated webpage) without direct access to memory.
- The attack enables cross-session, cross-site compromise and can bypass permission-based defenses because the contamination is silently activated during future tasks.
- Experiments on (Visual)WebArena report substantial attack success rates (up to 32.5% on GPT-5-mini, 23.4% on GPT-5.2, and 19.5% on GPT-OSS-120B), indicating the threat is practical rather than purely theoretical.
- A key factor is “Frustration Exploitation,” where agent stress (dropped clicks/garbled text) increases vulnerability up to 8×, and the authors find that stronger models are not necessarily more secure.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat Asia
AI Business

How Bash Command Safety Analysis Works in AI Systems
Dev.to

How I Built an AI Agent That Earns USDC While I Sleep — A Complete Guide
Dev.to

How to Get Better Output from AI Tools (Without Burning Time and Tokens)
Dev.to

How I Added LangChain4j Without Letting It Take Over My Spring Boot App
Dev.to