Greedy Is a Strong Default: Agents as Iterative Optimizers
arXiv cs.AI / 3/31/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper replaces classical random candidate proposal steps with an LLM agent that uses evaluation diagnostics to propose better candidates in an iterative optimization loop.
- Experiments across four discrete, mixed, and continuous optimization tasks show that greedy hill climbing with early stopping matches or outperforms more complex configurations while using substantially fewer evaluations.
- A cross-task ablation finds that simulated annealing, parallel investigators, and using a second LLM model (OpenAI Codex) do not improve outcomes and instead increase evaluation cost by about 2–3×.
- Results indicate the LLM’s learned prior is strong enough that sophisticated acceptance rules add limited value, with round 1 often accounting for most of the gains.
- Beyond performance, the approach can yield interpretable outputs, such as cancer classification rules that reflect established cytopathology concepts.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.



