ElephantBroker: A Knowledge-Grounded Cognitive Runtime for Trustworthy AI Agents

arXiv cs.AI / 3/27/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • ElephantBroker is an open-source “cognitive runtime” that combines a Neo4j knowledge graph with a Qdrant vector store to give LLM agents durable, verifiable memory with tracked provenance and trustworthiness.
  • It implements a full cognitive loop (store, retrieve, score, compose, protect, learn) using a hybrid multi-source retrieval pipeline, an evidence verification model, and goal-aware context assembly designed for context-budget constraints.
  • The system adds layered safety controls including guard pipelines, an AI firewall for enforceable tool-call interception, and multi-tier safety scanning to support safer agent behavior in high-stakes multi-turn settings.
  • ElephantBroker includes a consolidation engine and an authority model for multi-organization identity with hierarchical access control, plus continuous compaction to manage memory quality over time.
  • The authors report architectural validation via a test suite of 2,200+ unit/integration/end-to-end tests and describe modular deployments (from lightweight to enterprise-grade) with management dashboards for human oversight.

Abstract

Large Language Model based agents increasingly operate in high stakes, multi turn settings where factual grounding is critical, yet their memory systems typically rely on flat key value stores or plain vector retrieval with no mechanism to track the provenance or trustworthiness of stored knowledge. We present ElephantBroker, an open source cognitive runtime that unifies a Neo4j knowledge graph with a Qdrant vector store through the Cognee SDK to provide durable, verifiable agent memory. The system implements a complete cognitive loop (store, retrieve, score, compose, protect, learn) comprising a hybrid five source retrieval pipeline, an eleven dimension competitive scoring engine for budget constrained context assembly, a four state evidence verification model, a five stage context lifecycle with goal aware assembly and continuous compaction, a six layer cheap first guard pipeline for safety enforcement, an AI firewall providing enforceable tool call interception and multi tier safety scanning, a nine stage consolidation engine that strengthens useful patterns while decaying noise, and a numeric authority model governing multi organization identity with hierarchical access control. Architectural validation through a comprehensive test suite of over 2,200 tests spanning unit, integration, and end to end levels confirms subsystem correctness. The modular design supports three deployment tiers, five profile presets with inheritance, multi gateway isolation, and a management dashboard for human oversight, enabling configurations from lightweight memory only agents to full cognitive runtimes with enterprise grade safety and auditability.