From Experience to Skill: Multi-Agent Generative Engine Optimization via Reusable Strategy Learning

arXiv cs.AI / 4/22/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research

Key Points

  • The paper argues that existing Generative Engine Optimization (GEO) approaches optimize each query in isolation and cannot reuse or transfer effective optimization strategies across tasks and engines.
  • It reframes GEO as a strategy learning problem and proposes MAGEO, a multi-agent framework that uses coordinated planning, editing, and fidelity-aware evaluation to distill reusable, engine-specific “optimization skills.”
  • To support controlled evaluation and causal attribution of changes, the authors introduce the Twin Branch Evaluation Protocol and the DSV-CF metric, which combines semantic visibility with attribution accuracy.
  • They release MSME-GEO-Bench, a benchmark covering multiple scenarios and multiple engines based on real-world user queries, and show that MAGEO improves both content visibility and citation fidelity over heuristic baselines across three mainstream engines.
  • Ablation studies indicate that engine-specific preference modeling and strategy reuse are key contributors to the performance gains, pointing to a scalable, learning-driven paradigm for more trustworthy GEO.

Abstract

Generative engines (GEs) are reshaping information access by replacing ranked links with citation-grounded answers, yet current Generative Engine Optimization (GEO) methods optimize each instance in isolation, unable to accumulate or transfer effective strategies across tasks and engines. We reframe GEO as a strategy learning problem and propose MAGEO, a multi-agent framework in which coordinated planning, editing, and fidelity-aware evaluation serve as the execution layer, while validated editing patterns are progressively distilled into reusable, engine-specific optimization skills. To enable controlled assessment, we introduce a Twin Branch Evaluation Protocol for causal attribution of content edits and DSV-CF, a dual-axis metric that unifies semantic visibility with attribution accuracy. We further release MSME-GEO-Bench, a multi-scenario, multi-engine benchmark grounded in real-world queries. Experiments on three mainstream engines show that MAGEO substantially outperforms heuristic baselines in both visibility and citation fidelity, with ablations confirming that engine-specific preference modeling and strategy reuse are central to these gains, suggesting a scalable learning-driven paradigm for trustworthy GEO. Code is available at https://github.com/Wu-beining/MAGEO