AI Navigate

PRECEPT: Planning Resilience via Experience, Context Engineering & Probing Trajectories A Unified Framework for Test-Time Adaptation with Compositional Rule Learning and Pareto-Guided Prompt Evolution

arXiv cs.AI / 3/11/2026

Ideas & Deep AnalysisModels & Research

Key Points

  • PRECEPT is a unified framework designed for test-time adaptation of large language model agents that addresses retrieval degradation, rule composition reliability, and detection of stale or adversarial knowledge.
  • The framework integrates three components: deterministic exact-match rule retrieval, conflict-aware memory with Bayesian reliability and rule invalidation, and COMPASS, a Pareto-guided prompt evolution loop.
  • PRECEPT provides significant improvements including a +41.1pp first-try performance advantage, +33.3pp compositional generalization, and robust adversarial static knowledge handling.
  • The approach eliminates partial-match retrieval errors and supports compositional stacking via a semantic hierarchy, enhancing continuous learning and drift recovery.
  • Experimental results demonstrate statistically significant gains in efficiency and robustness compared to existing methods, with reduced steps and strong recovery from adversarial and drift challenges.

Abstract

LLM agents that store knowledge as natural language suffer steep retrieval degradation as condition count grows, often struggle to compose learned rules reliably, and typically lack explicit mechanisms to detect stale or adversarial knowledge. We introduce PRECEPT, a unified framework for test-time adaptation with three tightly coupled components: (1) deterministic exact-match rule retrieval over structured condition keys, (2) conflict-aware memory with Bayesian source reliability and threshold-based rule invalidation, and (3) COMPASS, a Pareto-guided prompt-evolution outer loop. Exact retrieval eliminates partial-match interpretation errors on the deterministic path (0% by construction, vs 94.4% under Theorem~B.6's independence model at N=10) and supports compositional stacking through a semantic tier hierarchy; conflict-aware memory resolves static--dynamic disagreements and supports drift adaptation; COMPASS evaluates prompts through the same end-to-end execution pipeline. Results (9--10 seeds): PRECEPT achieves a +41.1pp first-try advantage over Full Reflexion (d>1.9), +33.3pp compositional generalization (d=1.55), 100% P_1 on 2-way logistics compositions (d=2.64), +40--55pp continuous learning gains, strong eventual robustness under adversarial static knowledge (100% logistics with adversarial SK active; partial recovery on integration), +55.0pp drift recovery (d=0.95, p=0.031), and 61% fewer steps. Core comparisons are statistically significant, often at p<0.001.