GenericAgent: A Token-Efficient Self-Evolving LLM Agent via Contextual Information Density Maximization (V1.0)

arXiv cs.CL / 4/21/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper argues that long-horizon LLM agent performance depends less on raw context length and more on maintaining decision-relevant information within a limited context budget.
  • It introduces GenericAgent (GA), a general-purpose self-evolving agent system built on “context information density maximization” to prevent important details from being pushed out.
  • GA combines a minimal atomic tool set, a hierarchical on-demand memory with a small default view, and a self-evolution mechanism that converts verified past trajectories into reusable SOPs and executable code.
  • A context truncation and compression layer is used to preserve information density during long runs, improving efficiency in tool use, memory effectiveness, and execution.
  • Experiments reported in the abstract claim GA outperforms leading agent systems across multiple criteria (task completion, tool efficiency, memory, self-evolution, and web browsing) while using fewer tokens and interactions, and it continues evolving over time.

Abstract

Long-horizon large language model (LLM) agents are fundamentally limited by context. As interactions become longer, tool descriptions, retrieved memories, and raw environmental feedback accumulate and push out the information needed for decision-making. At the same time, useful experience gained from tasks is often lost across episodes. We argue that long-horizon performance is determined not by context length, but by how much decision-relevant information is maintained within a finite context budget. We present GenericAgent (GA), a general-purpose, self-evolving LLM agent system built around a single principle: context information density maximization. GA implements this through four closely connected components: a minimal atomic tool set that keeps the interface simple, a hierarchical on-demand memory that only shows a small high-level view by default, a self-evolution mechanism that turns verified past trajectories into reusable SOPs and executable code, and a context truncation and compression layer that maintains information density during long executions. Across task completion, tool use efficiency, memory effectiveness, self-evolution, and web browsing, GA consistently outperforms leading agent systems while using significantly fewer tokens and interactions, and it continues to evolve over time. Project: https://github.com/lsdefine/GenericAgent