AEL: Agent Evolving Learning for Open-Ended Environments

arXiv cs.CL / 4/24/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that LLM agents in open-ended, long-horizon environments are currently mostly stateless, and the key challenge is figuring out how to *use* stored experience rather than what to remember.
  • It introduces Agent Evolving Learning (AEL), a two-timescale approach where a fast Thompson-sampling bandit selects a memory retrieval policy each episode while a slower LLM reflection module diagnoses failure patterns and updates the agent’s prompt with causal insights.
  • In a sequential portfolio benchmark (10 diverse tickers, 208 episodes, 5 seeds), AEL achieves a Sharpe ratio of 2.13±0.47, outperforming five prior self-improving methods and all non-LLM baselines while showing the lowest variance among LLM-based approaches.
  • An ablation study across nine variants finds that combining memory and reflection yields a 58% cumulative improvement over a stateless baseline, but adding other mechanisms (e.g., planner evolution, tool selection, cold-start initialization, skill extraction, and different credit assignment methods) consistently degrades performance.
  • The results suggest that the bottleneck in agent self-improvement is self-diagnosing how to interpret and apply experience, and that increasing architectural complexity may hurt rather than help.

Abstract

LLM agents increasingly operate in open-ended environments spanning hundreds of sequential episodes, yet they remain largely stateless: each task is solved from scratch without converting past experience into better future behavior. The central obstacle is not \emph{what} to remember but \emph{how to use} what has been remembered, including which retrieval policy to apply, how to interpret prior outcomes, and when the current strategy itself must change. We introduce \emph{Agent Evolving Learning} (\ael{}), a two-timescale framework that addresses this obstacle. At the fast timescale, a Thompson Sampling bandit learns which memory retrieval policy to apply at each episode; at the slow timescale, LLM-driven reflection diagnoses failure patterns and injects causal insights into the agent's decision prompt, giving it an interpretive frame for the evidence it retrieves. On a sequential portfolio benchmark (10 sector-diverse tickers, 208 episodes, 5 random seeds), \ael{} achieves a Sharpe ratio of 2.13\pm0.47, outperforming five published self-improving methods and all non-LLM baselines while maintaining the lowest variance among all LLM-based approaches. A nine-variant ablation reveals a ``less is more'' pattern: memory and reflection together produce a 58\% cumulative improvement over the stateless baseline, yet every additional mechanism we test (planner evolution, per-tool selection, cold-start initialization, skill extraction, and three credit assignment methods) \emph{degrades} performance. This demonstrates that the bottleneck in agent self-improvement is \emph{self-diagnosing how to use} experience rather than adding architectural complexity. Code and data: https://github.com/WujiangXu/AEL.