Understanding the Challenges in Iterative Generative Optimization with LLMs

arXiv cs.LG / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies generative optimization with LLMs, where a model iteratively improves artifacts using execution feedback, but argues the approach is often brittle in practice.
  • It explains that brittleness stems from “hidden” design choices required to build a learning loop, including what the optimizer is allowed to edit and what constitutes the correct learning evidence at each update.
  • The authors investigate three key application factors—starting artifact, the credit horizon over execution traces, and how trials/errors are batched into learning evidence—and show they strongly affect outcomes.
  • Case studies across MLAgentBench, Atari, and BigBench Extra Hard indicate that these choices determine whether optimization succeeds, and that effects are non-monotonic (e.g., larger minibatches do not always improve generalization).
  • The work concludes there is no simple universal recipe for setting up learning loops across domains and provides practical guidance to make these decisions explicit for productionization.

Abstract

Generative optimization uses large language models (LLMs) to iteratively improve artifacts (such as code, workflows or prompts) using execution feedback. It is a promising approach to building self-improving agents, yet in practice remains brittle: despite active research, only 9% of surveyed agents used any automated optimization. We argue that this brittleness arises because, to set up a learning loop, an engineer must make ``hidden'' design choices: What can the optimizer edit and what is the "right" learning evidence to provide at each update? We investigate three factors that affect most applications: the starting artifact, the credit horizon for execution traces, and batching trials and errors into learning evidence. Through case studies in MLAgentBench, Atari, and BigBench Extra Hard, we find that these design decisions can determine whether generative optimization succeeds, yet they are rarely made explicit in prior work. Different starting artifacts determine which solutions are reachable in MLAgentBench, truncated traces can still improve Atari agents, and larger minibatches do not monotonically improve generalization on BBEH. We conclude that the lack of a simple, universal way to set up learning loops across domains is a major hurdle for productionization and adoption. We provide practical guidance for making these choices.

Understanding the Challenges in Iterative Generative Optimization with LLMs | AI Navigate