Provable Accelerated Bayesian Optimization with Knowledge Transfer

arXiv stat.ML / 4/29/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses accelerating Bayesian optimization on a target task by transferring historical information from related source tasks.
  • It introduces DeltaBO, which quantifies uncertainty using the difference function between source and target tasks and supports cases where they lie in different RKHS settings.
  • The authors provide theoretical regret bounds for DeltaBO, achieving O~(sqrt(T (T/N + gamma_delta))) under mild assumptions, improving over prior transfer approaches that lacked guarantees or matched non-transfer performance.
  • Empirical results on real hyperparameter-tuning benchmarks and synthetic experiments show DeltaBO outperforms baselines and aligns with the proposed theory.
  • The work also releases an implementation on GitHub for practical evaluation and reuse.

Abstract

We study how to accelerate Bayesian optimization (BO) on a target task by transferring historical knowledge from related source tasks. Existing work on BO with knowledge transfer either lacks theoretical guarantees or achieves the same regret as BO in the non-transfer setting, \widetilde{O}(\sqrt{T \gamma_f}), where T is the number of evaluations of the target function and \gamma_f denotes its information gain. In this paper, we propose the DeltaBO algorithm, which builds a novel uncertainty-quantification approach on the difference function \delta between the source and target functions, which are allowed to belong to different Reproducing Kernel Hilbert Spaces (RKHSs). Under mild assumptions, we prove that the regret of DeltaBO is of order \widetilde{O}(\sqrt{T (T/N + \gamma_\delta)}), where N denotes the number of evaluations from source tasks and typically N \gg T. In many applications, source and target tasks are similar, which implies that \gamma_\delta can be much smaller than \gamma_f. Empirical studies on both real-world hyperparameter-tuning tasks and synthetic functions show that DeltaBO outperforms other baseline methods and also verify our theoretical claims. Our code is available on GitHub.