Learning to Evolve: A Self-Improving Framework for Multi-Agent Systems via Textual Parameter Graph Optimization

arXiv cs.AI / 4/23/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes Textual Parameter Graph Optimization (TPGO) to automatically optimize multi-agent systems with structural awareness, overcoming limitations of flat prompt tuning methods.
  • TPGO represents a multi-agent system as a Textual Parameter Graph (TPG) with modular, optimizable nodes such as agents, tools, and workflows.
  • It introduces “textual gradients,” i.e., natural-language guidance generated from execution traces, to diagnose failures and recommend fine-grained changes.
  • The framework’s core, Group Relative Agent Optimization (GRAO), uses meta-learning from past optimization successes and failures so the system can improve its own optimization strategy over time.
  • Experiments on benchmarks including GAIA and MCP-Universe indicate that TPGO substantially boosts the success rates of leading agent frameworks through automated self-improving optimization.

Abstract

Designing and optimizing multi-agent systems (MAS) is a complex, labor-intensive process of "Agent Engineering." Existing automatic optimization methods, primarily focused on flat prompt tuning, lack the structural awareness to debug the intricate web of interactions in MAS. More critically, these optimizers are static; they do not learn from experience to improve their own optimization strategies. To address these gaps, we introduce Textual Parameter Graph Optimization (TPGO), a framework that enables a multi-agent system to learn to evolve. TPGO first models the MAS as a Textual Parameter Graph (TPG), where agents, tools, and workflows are modular, optimizable nodes. To guide evolution, we derive "textual gradients," structured natural language feedback from execution traces, to pinpoint failures and suggest granular modifications. The core of our framework is Group Relative Agent Optimization (GRAO), a novel meta-learning strategy that learns from historical optimization experiences. By analyzing past successes and failures, GRAO becomes progressively better at proposing effective updates, allowing the system to learn how to optimize itself. Extensive experiments on complex benchmarks like GAIA and MCP-Universe show that TPGO significantly enhances the performance of state-of-the-art agent frameworks, achieving higher success rates through automated, self-improving optimization.