SuperLocalMemory V3.3: The Living Brain -- Biologically-Inspired Forgetting, Cognitive Quantization, and Multi-Channel Retrieval for Zero-LLM Agent Memory Systems

arXiv cs.AI / 4/7/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • SuperLocalMemory V3.3 (“The Living Brain”) is presented as a local-first agent memory system for coding agents, aiming to overcome limitations of single-channel vector-database memory and reliance on cloud LLMs.
  • The work introduces Fisher-Rao Quantization-Aware Distance (FRQAD) to better prefer high-fidelity embeddings over quantized ones, alongside a mathematically defined, lifecycle-aware forgetting mechanism inspired by Ebbinghaus.
  • It proposes a 7-channel cognitive retrieval approach (including semantic, keyword, entity graph, temporal, spreading activation, consolidation, and Hopfield associative channels) and reports improved zero-LLM performance on LoCoMo Mode A.
  • The system also implements long-term implicit memory via soft prompts and includes an automated “auto-cognitive pipeline” to manage the memory lifecycle end-to-end.
  • V3.3 is released as open source (Elastic License 2.0), runs entirely on CPU, and the paper reports results such as 70.4% on LoCoMo Mode A in zero-LLM mode and gains on multi-hop and adversarial settings.

Abstract

AI coding agents operate in a paradox: they possess vast parametric knowledge yet cannot remember a conversation from an hour ago. Existing memory systems store text in vector databases with single-channel retrieval, require cloud LLMs for core operations, and implement none of the cognitive processes that make human memory effective. We present SuperLocalMemory V3.3 ("The Living Brain"), a local-first agent memory system implementing the full cognitive memory taxonomy with mathematical lifecycle dynamics. Building on the information-geometric foundations of V3.2 (arXiv:2603.14588), we introduce five contributions: (1) Fisher-Rao Quantization-Aware Distance (FRQAD) -- a new metric on the Gaussian statistical manifold achieving 100% precision at preferring high-fidelity embeddings over quantized ones (vs 85.6% for cosine), with zero prior art; (2) Ebbinghaus Adaptive Forgetting with lifecycle-aware quantization -- the first mathematical forgetting curve in local agent memory coupled to progressive embedding compression, achieving 6.7x discriminative power; (3) 7-channel cognitive retrieval spanning semantic, keyword, entity graph, temporal, spreading activation, consolidation, and Hopfield associative channels, achieving 70.4% on LoCoMo in zero-LLM Mode A; (4) memory parameterization implementing Long-Term Implicit memory via soft prompts; (5) zero-friction auto-cognitive pipeline automating the complete memory lifecycle. On LoCoMo, V3.3 achieves 70.4% in Mode A (zero-LLM), with +23.8pp on multi-hop and +12.7pp on adversarial. V3.2 achieved 74.8% Mode A and 87.7% Mode C; the 4.4pp gap reflects a deliberate architectural trade-off. SLM V3.3 is open source under the Elastic License 2.0, runs entirely on CPU, with over 5,000 monthly downloads.