Context Cartography: Toward Structured Governance of Contextual Space in Large Language Model Systems

arXiv cs.AI / 2026/3/24

💬 オピニオンSignals & Early TrendsIdeas & Deep AnalysisModels & Research

要点

  • The paper argues that simply increasing LLM context window size is not sufficient because transformer “contextual space” has structural gradients, salience asymmetries, and degradation effects over long distances (e.g., “lost in the middle”).
  • It proposes “Context Cartography,” a formal governance framework that partitions informational context into three zones—black fog (unobserved), gray fog (stored memory), and visible field (active reasoning surface)—and defines seven operators to manage transitions across and within these zones.
  • The seven cartographic operators (reconnaissance, selection, simplification, aggregation, projection, displacement, and layering) are organized by transformation type and zone scope, with derivations based on coverage analysis of non-trivial zone transformations.
  • The framework is grounded in transformer attention salience geometry, framing the operators as compensations for issues like linear prefix memory, append-only state, and entropy accumulation as context grows.
  • Using analysis of four contemporary systems (Claude Code, Letta, MemOS, and OpenViking), the authors claim independent convergence of these operators across the industry and provide testable, benchmarkable ablation/diagnostic predictions.

Abstract

The prevailing approach to improving large language model (LLM) reasoning has centered on expanding context windows, implicitly assuming that more tokens yield better performance. However, empirical evidence - including the "lost in the middle" effect and long-distance relational degradation - demonstrates that contextual space exhibits structural gradients, salience asymmetries, and entropy accumulation under transformer architectures. We introduce Context Cartography, a formal framework for the deliberate governance of contextual space. We define a tripartite zonal model partitioning the informational universe into black fog (unobserved), gray fog (stored memory), and the visible field (active reasoning surface), and formalize seven cartographic operators - reconnaissance, selection, simplification, aggregation, projection, displacement, and layering - as transformations governing information transitions between and within zones. The operators are derived from a systematic coverage analysis of all non-trivial zone transformations and are organized by transformation type (what the operator does) and zone scope (where it applies). We ground the framework in the salience geometry of transformer attention, characterizing cartographic operators as necessary compensations for linear prefix memory, append-only state, and entropy accumulation under expanding context. An analysis of four contemporary systems (Claude Code, Letta, MemOS, and OpenViking) provides interpretive evidence that these operators are converging independently across the industry. We derive testable predictions from the framework - including operator-specific ablation hypotheses - and propose a diagnostic benchmark for empirical validation.