Neural Global Optimization via Iterative Refinement from Noisy Samples

arXiv cs.LG / 4/7/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a neural method for global optimization of noisy black-box functions that aims to avoid local minima issues common to approaches like Bayesian Optimization.
  • The model uses noisy function samples plus a spline representation to iteratively refine an initial guess toward the true global minimum, without requiring gradient information or multiple restarts.
  • It is trained on synthetic randomly generated functions with known global minima (via exhaustive search), and tested on multi-modal benchmarks.
  • Experimental results show a mean error of 8.05% versus 36.24% for spline initialization, along with 72% of test cases reaching <10% error, suggesting the system learns optimization behavior rather than only fitting curves.
  • The architecture incorporates multiple input modalities (function values, derivatives, and spline coefficients) alongside iterative position updates to improve robustness across challenging landscapes.

Abstract

Global optimization of black-box functions from noisy samples is a fundamental challenge in machine learning and scientific computing. Traditional methods such as Bayesian Optimization often converge to local minima on multi-modal functions, while gradient-free methods require many function evaluations. We present a novel neural approach that learns to find global minima through iterative refinement. Our model takes noisy function samples and their fitted spline representation as input, then iteratively refines an initial guess toward the true global minimum. Trained on randomly generated functions with ground truth global minima obtained via exhaustive search, our method achieves a mean error of 8.05 percent on challenging multi-modal test functions, compared to 36.24 percent for the spline initialization, a 28.18 percent improvement. The model successfully finds global minima in 72 percent of test cases with error below 10 percent, demonstrating learned optimization principles rather than mere curve fitting. Our architecture combines encoding of multiple modalities including function values, derivatives, and spline coefficients with iterative position updates, enabling robust global optimization without requiring derivative information or multiple restarts.