ADEMA: A Knowledge-State Orchestration Architecture for Long-Horizon Knowledge Synthesis with LLMAgents

arXiv cs.AI / 4/29/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces ADEMA, a knowledge-state orchestration architecture designed to make long-horizon LLM knowledge synthesis more reliable by preventing knowledge-state drift and preserving evidence continuity across rounds.
  • ADEMA emphasizes explicit epistemic bookkeeping, dual heterogeneous evaluator governance, adaptive task-mode switching, reputation-shaped resource allocation, and checkpoint-resumable persistence to improve trajectory discipline and cost-quality behavior.
  • It also uses segment-level memory condensation, artifact-first assembly, and final-validity checking with a safe fallback rather than relying solely on generic multi-agent runtime assumptions.
  • Experiments using an 60-run fixed mechanism matrix show that removing checkpoint/resume caused the only invalid run, especially under interruption-sensitive resume conditions, highlighting recoverable continuity as a key driver.
  • The authors conclude that dual evaluation, segment synthesis, and dynamic governance function mainly as supporting control mechanisms that shape progression and outcomes, not as universal prerequisites for success.

Abstract

Long-horizon LLM tasks often fail not because a single answer is unattainable, but because knowledge states drift across rounds, intermediate commitments remain implicit, and interruption fractures the evolving evidence chain. This paper presents ADEMA as a knowledge-state orchestration architecture for long-horizon knowledge synthesis rather than as a generic multi-agent runtime. The architecture combines explicit epistemic bookkeeping, heterogeneous dual-evaluator governance, adaptive task-mode switching, reputation-shaped resource allocation, checkpoint-resumable persistence, segment-level memory condensation, artifact-first assembly, and final-validity checking with safe fallback. Evidence is drawn entirely from existing materials: a four-scenario showcase package, a fixed 60-run mechanism matrix, targeted micro-ablation and artifact-chain supplements, and a repaired protocol-level benchmark in which code-oriented evaluation is the clearest quality-sensitive mechanism block. Across the fixed matrix, removing checkpoint/resume produced the only invalid run, and it did so in the interruption-sensitive resume condition. By contrast, dual evaluation, segment synthesis, and dynamic governance are best interpreted as supporting control mechanisms that shape trajectory discipline, explicit artifact progression, and cost-quality behavior rather than as universal binary prerequisites for completion. The contribution is therefore a knowledge-state orchestration architecture in which explicit epistemic state transition, evidence-bearing artifact progression, and recoverable continuity are the primary design commitments.