BEAM: Bi-level Memory-adaptive Algorithmic Evolution for LLM-Powered Heuristic Design

arXiv cs.AI / 4/15/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces BEAM (Bi-level Memory-adaptive Algorithmic Evolution) to improve LLM-based hyper-heuristic design beyond single-function optimization by framing it as bi-level optimization.
  • BEAM uses an exterior genetic algorithm layer to evolve high-level algorithmic structures with function placeholders, while an interior Monte Carlo Tree Search layer fills in those placeholders to realize candidate solvers.
  • An Adaptive Memory module is added to support more complex code generation during the heuristic design process.
  • To enable better evaluation and generation, the authors propose a Knowledge Augmentation (KA) pipeline and argue that starting from scratch or only from code templates limits LHH performance.
  • Experiments across several optimization problems show BEAM significantly better results than prior LHHs, including a 37.84% reduction in optimality gap for CVRP hybrid algorithm design and new performance on Maximum Independent Set (MIS) tasks versus KaMIS.

Abstract

Large Language Model-based Hyper Heuristic (LHH) has recently emerged as an efficient way for automatic heuristic design. However, most existing LHHs just perform well in optimizing a single function within a pre-defined solver. Their single-layer evolution makes them not effective enough to write a competent complete solver. While some variants incorporate hyperparameter tuning or attempt to generate complex code through iterative local modifications, they still lack a high-level algorithmic modeling, leading to limited exploration efficiency. To address this, we reformulate heuristic design as a Bi-level Optimization problem and propose \textbf{BEAM} (Bi-level Memory-adaptive Algorithmic Evolution). BEAM's exterior layer evolves high-level algorithmic structures with function placeholders through genetic algorithm (GA), while the interior layer realizes these placeholders via Monte Carlo Tree Search (MCTS). We further introduce an Adaptive Memory module to facilitate complex code generation. To support the evaluation for complex code generation, we point out the limitations of starting LHHs from scratch or from code templates and introduce a Knowledge Augmentation (KA) Pipeline. Experimental results on several optimization problems demonstrate that BEAM significantly outperforms existing LHHs, notably reducing the optimality gap by 37.84\% on aggregate in CVRP hybrid algorithm design. BEAM also designs a heuristic that outperforms SOTA Maximum Independent Set (MIS) solver KaMIS.