ATANT: An Evaluation Framework for AI Continuity
arXiv cs.AI / 4/10/2026
📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisTools & Practical UsageModels & Research
Key Points
- The paper introduces ATANT, an open, system-agnostic evaluation framework that measures “AI continuity” (persistence, updating, disambiguation, and reconstruction of meaningful context over time) rather than just using memory components like RAG or long-context windows.
- Continuity is defined via seven required properties, alongside a 10-checkpoint evaluation methodology that can run without an LLM inside the evaluation loop to avoid evaluation-time bias.
- ATANT provides a narrative test corpus of 250 life-domain stories with 1,835 verification questions, enabling repeatable benchmarking across scenarios.
- A reference implementation is evaluated over multiple suite iterations, improving from 58% with a legacy architecture to 100% in isolated testing and achieving 96% at the 250-story cumulative scale, where cross-contamination is a key failure mode.
- The framework, example stories, and protocol are published on GitHub, with the full 250-story corpus planned for incremental release.


