Towards Fully Parameter-Free Stochastic Optimization: Grid Search with Self-Bounding Analysis

arXiv cs.LG / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies “fully parameter-free” stochastic optimization methods that do not require any unverifiable assumptions about the true problem parameters.
  • It introduces Grasp, a general grid-search framework that uses a novel self-bounding analysis to automatically determine parameter search ranges, avoiding reliance on known bounds.
  • The authors show the approach works broadly, including non-convex optimization with near-optimal convergence rates (up to logarithmic factors).
  • In the convex setting, the proposed parameter-free methods are competitive in both acceleration and universality.
  • The work also provides a sharper theoretical guarantee for the final model ensemble stage of the grid-search framework under interpolated variance characterization.

Abstract

Parameter-free stochastic optimization aims to design algorithms that are agnostic to the underlying problem parameters while still achieving convergence rates competitive with optimally tuned methods. While some parameter-free methods do not require the specific values of the problem parameters, they still rely on prior knowledge, such as the lower or upper bounds of them. We refer to such methods as ``partially parameter-free''. In this work, we target achieving ``fully parameter-free'' methods, i.e., the algorithmic inputs do not need to satisfy any unverifiable condition related to the true problem parameters. We propose a powerful and general grid search framework, named \textsc{Grasp}, with a novel self-bounding analysis technique that effectively determines the search ranges of parameters, in contrast to previous work. Our method demonstrates generality in: (i) the non-convex case, where we propose a fully parameter-free method that achieves near-optimal convergence rate, up to logarithmic factors; (ii) the convex case, where our parameter-free methods are competitive with strong performance in terms of acceleration and universality. Finally, we contribute a sharper guarantee for the model ensemble, a final step of the grid search framework, under interpolated variance characterization.