AgenticRS-EnsNAS: Ensemble-Decoupled Self-Evolving Architecture Search

arXiv cs.LG / 3/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Ensemble-Decoupled Architecture Search (EDAS), a framework that predicts system-level performance from single-learner evaluation to reduce per-candidate search cost from O(M) to O(1) while maintaining O(M) deployment cost for validated winners.
  • It provides a sufficient condition for monotonic ensemble improvement under homogeneity: rho(pi) < rho(pi_old) - (M/(M-1)) * (Delta E(pi)/sigma^2(pi)), with Delta E, rho, and sigma^2 estimable from lightweight dual-learner training.
  • The approach decouples architecture search from full ensemble training and outlines three solution strategies: closed-form optimization for tractable continuous pi, constrained differentiable optimization for intractable continuous pi, and LLM-driven search with iterative monotonic acceptance for discrete pi.
  • The framework highlights two improvement mechanisms—base diversity gain and accuracy gain—providing actionable NAS design principles, with rigorous theoretical results and planned empirical validation in the journal extension.

Abstract

Neural Architecture Search (NAS) deployment in industrial production systems faces a fundamental validation bottleneck: verifying a single candidate architecture pi requires evaluating the deployed ensemble of M models, incurring prohibitive O(M) computational cost per candidate. This cost barrier severely limits architecture iteration frequency in real-world applications where ensembles (M=50-200) are standard for robustness. This work introduces Ensemble-Decoupled Architecture Search, a framework that leverages ensemble theory to predict system-level performance from single-learner evaluation. We establish the Ensemble-Decoupled Theory with a sufficient condition for monotonic ensemble improvement under homogeneity assumptions: a candidate architecture pi yields lower ensemble error than the current baseline if rho(pi) < rho(pi_old) - (M / (M - 1)) * (Delta E(pi) / sigma^2(pi)), where Delta E, rho, and sigma^2 are estimable from lightweight dual-learner training. This decouples architecture search from full ensemble training, reducing per-candidate search cost from O(M) to O(1) while maintaining O(M) deployment cost only for validated winners. We unify solution strategies across pipeline continuity: (1) closed-form optimization for tractable continuous pi (exemplified by feature bagging in CTR prediction), (2) constrained differentiable optimization for intractable continuous pi, and (3) LLM-driven search with iterative monotonic acceptance for discrete pi. The framework reveals two orthogonal improvement mechanisms -- base diversity gain and accuracy gain -- providing actionable design principles for industrial-scale NAS. All theoretical derivations are rigorous with detailed proofs deferred to the appendix. Comprehensive empirical validation will be included in the journal extension of this work.