Learning to Forget: Sleep-Inspired Memory Consolidation for Resolving Proactive Interference in Large Language Models
arXiv cs.AI / 3/17/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- SleepGate proposes a sleep-like inference cycle to mitigate proactive interference by managing the KV cache in transformer-based LLMs.
- It introduces three components: a conflict-aware temporal tagger, a lightweight forgetting gate, and a consolidation module that evict/compress stale entries and summarize the remaining ones.
- Inference sleep cycles are governed by an adaptive entropy-based trigger and trained with a dual-phase objective for wake-language modeling and post-consolidation retrieval.
- Theoretical analysis shows the interference horizon reduces from O(n) to O(log n), improving retrieval as the context length grows.
- Empirical results on a small 4-layer transformer (793K parameters) show 99.5% retrieval accuracy at PI depth 5 and 97.0% at depth 10, outperforming baselines and suggesting an architecture-level solution beyond prompt engineering.




