AI Navigate

Selective Memory for Artificial Intelligence: Write-Time Gating with Hierarchical Archiving

arXiv cs.AI / 3/18/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes write-time gating with hierarchical archiving that filters incoming knowledge objects using composite salience scores (source reputation, novelty, reliability) and preserves version chains to retain prior states.
  • This approach addresses limitations of retrieval-augmented generation and parametric memory by avoiding indiscriminate storage and enabling selective updates.
  • In evaluations with real LLMs and without oracle quality labels, write-time gating achieves 100 percent accuracy, compared with 13 percent for ungated stores.
  • Under distractor scaling (8:1), read-time filtering collapses to 0 percent while write gating maintains 100 percent, and the method reduces query-time cost to one-ninth of Self-RAG, with accuracy gains observed across Wikipedia, pharmacology data, and post-cutoff arXiv papers.

Abstract

Retrieval-augmented generation stores all content indiscriminately, degrading accuracy as noise accumulates. Parametric approaches compress knowledge into weights, precluding selective updates. Neither mirrors biological memory, which gates encoding based on salience and archives rather than deletes superseded information. We introduce write-time gating that filters incoming knowledge objects using composite salience scores (source reputation, novelty, reliability) while maintaining version chains that preserve prior states. Using real LLM evaluation without oracle access to quality labels, write gating achieves 100 percent accuracy versus 13 percent for ungated stores. The critical finding emerges under distractor scaling: at 8:1 distractor ratios, read-time filtering (Self-RAG) collapses to 0 percent while write gating maintains 100 percent, revealing a structural advantage of write-time over read-time curation. Validation on Wikipedia (20 entities), procedurally generated pharmacology data, and 2026 arXiv papers confirms these findings. The gating advantage scales inversely with parametric memory support: +25pp for Wikipedia, +48pp for post-cutoff arXiv, +65pp for procedural data with zero training knowledge. Signal ablation confirms the method does not depend on oracle-correlated metadata. Write gating matches Self-RAG accuracy at one-ninth the query-time cost.