LLM as a Tool, Not an Agent: Code-Mined Tree Transformations for Neural Architecture Search

arXiv cs.LG / 4/21/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper introduces LLMasTool, a hierarchical, tree-based Neural Architecture Search (NAS) framework that uses LLMs as a tool to support model evolution rather than fully agentic code generation.
  • It automatically extracts reusable neural modules from arbitrary source code and represents candidate architectures as hierarchical trees, so evolution happens via reliable tree transformations instead of generating entire architectures as raw code.
  • The approach uses diversity-guided, Bayesian modeling at the coarse-planning level to improve exploration efficiency, while the LLM handles remaining design degrees of freedom to produce executable architectures.
  • By shifting from fully agentic LLM proposals toward algorithmic tree transformations, the method aims to reduce over-reliance on patterns biased toward the LLM’s training data.
  • Experiments show improved NAS performance versus existing methods, with gains of 0.69 (CIFAR-10), 1.83 (CIFAR-100), and 2.68 points (ImageNet16-120).

Abstract

Neural Architecture Search (NAS) aims to automatically discover high-performing deep neural network (DNN) architectures. However, conventional algorithm-driven NAS relies on carefully hand-crafted search spaces to ensure executability, which restricts open-ended exploration. Recent coding-based agentic approaches using large language models (LLMs) reduce manual design, but current LLMs struggle to reliably generate complex, valid architectures, and their proposals are often biased toward a narrow set of patterns observed in their training data. To bridge reliable algorithmic search with powerful LLM assistance, we propose LLMasTool, a hierarchical tree-based NAS framework for stable and open-ended model evolution. Our method automatically extracts reusable modules from arbitrary source code and represents full architectures as hierarchical trees, enabling evolution through reliable tree transformations rather than code generation. At each evolution step, coarse-level planning is governed by a diversity-guided algorithm that leverages Bayesian modeling to improve exploration efficiency, while the LLM resolves the remaining degrees of freedom to ensure a meaningful evolutionary trajectory and an executable generated architecture. With this formulation, instead of fully agentic LLM approaches, our method explores diverse directions beyond the inherent biases in the LLM. Our method improves over existing NAS methods by 0.69, 1.83, and 2.68 points on CIFAR-10, CIFAR-100, and ImageNet16-120, demonstrating its effectiveness.