Cognis: Context-Aware Memory for Conversational AI Agents
arXiv cs.CL / 4/23/2026
📰 NewsDeveloper Stack & InfrastructureTools & Practical UsageModels & Research
Key Points
- The paper introduces Lyzr Cognis, a unified, context-aware memory architecture designed to give conversational LLM agents persistent, cross-session memory for better long-term personalization.
- Cognis uses a multi-stage retrieval pipeline that combines OpenSearch BM25 keyword matching and Matryoshka-based vector similarity search, fused with Reciprocal Rank Fusion for more robust memory retrieval.
- A context-aware ingestion pipeline first retrieves existing memories before extraction, enabling intelligent version tracking to preserve full memory history while keeping the backend store consistent.
- The approach adds temporal boosting for time-sensitive queries and uses a BGE-2 cross-encoder reranker to improve the quality of the final retrieved results.
- Experiments on LoCoMo and LongMemEval across eight answer-generation models show state-of-the-art performance, and the system is open-source with production deployment for conversational AI applications.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat USA
AI Business

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Trajectory Forecasts in Unknown Environments Conditioned on Grid-Based Plans
Dev.to

Elevating Austria: Google invests in its first data center in the Alps.
Google Blog

10 AI Tools Every Developer Should Try in 2026
Dev.to