Beyond the Attention Stability Boundary: Agentic Self-Synthesizing Reasoning Protocols

arXiv cs.AI / 4/28/2026

📰 NewsModels & Research

Key Points

  • The paper identifies a systemic failure mode called the “Attention Latch” in decoder-only autoregressive Transformers, where historical context can overpower mid-task updates and anchor an agent to obsolete constraints.
  • It explains this behavior as a manifestation of “Information Over-squashing” and introduces a metacognitive approach, Self-Synthesizing Reasoning Protocols (SSRP), that separates high-level planning (Architect) from turn-by-turn execution (Executive).
  • Experiments on 9K trajectories using MultiWOZ 2.2 show that SSRP significantly outperforms stateless Vanilla ReAct baselines, locating an “Attention Stability Boundary” where baseline success collapses.
  • The authors validate a new metric, Aggregate Pivot Accuracy (APA), and test SSRP using multiple experimental tiers, including retrieval-based pilots and complex multi-fact synthesis tasks.
  • Across models including Gemini 3.1 Pro, Claude Sonnet 4.6, and DeepSeek V3.2, SSRP delivers large resilience gains, while audits also uncover a “Grounding Paradox” where highly stable models refuse to hallucinate under certain contamination conditions.

Abstract

As LLM agents transition to autonomous digital coworkers, maintaining deterministic goal-directedness in non-linear multi-turn conversations emerged as an architectural bottleneck. We identify and formalize a systemic failure mode termed the Attention Latch in decoder-only autoregressive Transformers. This phenomenon, a behavioral manifestation of Information Over-squashing, occurs when the cumulative probabilistic weight of historical context overrides mid-task updates, causing agents to remain anchored to obsolete constraints despite explicit contradictory instructions. We propose Self-Synthesizing Reasoning Protocols (SSRP), a metacognitive framework that implements a discrete separation between high-level architectural planning (Architect) and turn-by-turn procedural execution (Executive). We evaluate SSRP across 9K trajectories using the MultiWOZ 2.2 dataset and the Aggregate Pivot Accuracy (APA), a novel metric we validate by mapping its scores to the U-shaped 'Lost in the Middle' curve. We present 3 experimental tiers: a shallow recency-based retrieval pilot, a high-entropy SOP, and a semantic hijacked 3-hop Multi-Fact Synthesis task. Our results empirically locate the Attention Stability Boundary, where stateless Vanilla ReAct baselines for GPT 5.4 collapse to 0.1% success while SSRP achieves a 715X Resilience Lift. We demonstrate statistically significant gains across Gemini 3.1 Pro, Claude Sonnet 4.6 and DeepSeek V3.2. Audits confirm SSRP necessity by proving attentional lapse via a recursive reflexion baseline (100% success); decoupling the latch from positional bias through equidistant stress testing (90% accuracy); and formalizing SSRP via the Information Bottleneck principle and granularity ablations. Procedural Integrity audit (98.8% adherence) reveals a Grounding Paradox where high-stability models fail by refusing to hallucinate under retrieval-reasoning contamination.