PI is back! Switching Acquisition Functions in Bayesian Optimization
- URL: http://arxiv.org/abs/2211.01455v1
- Date: Wed, 2 Nov 2022 19:49:03 GMT
- Title: PI is back! Switching Acquisition Functions in Bayesian Optimization
- Authors: Carolin Benjamins, Elena Raponi, Anja Jankovic, Koen van der Blom,
Maria Laura Santoni, Marius Lindauer, and Carola Doerr
- Abstract summary: We benchmark SMAC3 using Expected Improvement (EI) and Probability of Improvement (PI) as acquisition functions.
One schedule aims to use EI's explorative behavior in the early optimization steps, and then switches to PI for a better exploitation in the final steps.
Our results suggest that a schedule that allocates the first 25 % of the optimization budget to EI and the last 75 % to PI is a reliable default.
- Score: 10.014619543479876
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Bayesian Optimization (BO) is a powerful, sample-efficient technique to
optimize expensive-to-evaluate functions. Each of the BO components, such as
the surrogate model, the acquisition function (AF), or the initial design, is
subject to a wide range of design choices. Selecting the right components for a
given optimization task is a challenging task, which can have significant
impact on the quality of the obtained results. In this work, we initiate the
analysis of which AF to favor for which optimization scenarios. To this end, we
benchmark SMAC3 using Expected Improvement (EI) and Probability of Improvement
(PI) as acquisition functions on the 24 BBOB functions of the COCO environment.
We compare their results with those of schedules switching between AFs. One
schedule aims to use EI's explorative behavior in the early optimization steps,
and then switches to PI for a better exploitation in the final steps. We also
compare this to a random schedule and round-robin selection of EI and PI. We
observe that dynamic schedules oftentimes outperform any single static one. Our
results suggest that a schedule that allocates the first 25 % of the
optimization budget to EI and the last 75 % to PI is a reliable default.
However, we also observe considerable performance differences for the 24
functions, suggesting that a per-instance allocation, possibly learned on the
fly, could offer significant improvement over the state-of-the-art BO designs.
Related papers
- Approximation-Aware Bayesian Optimization [34.56666383247348]
High-dimensional Bayesian optimization (BO) tasks often require 10,000 function evaluations before obtaining meaningful results.
We modify sparse variational Gaussian processes (SVGPs) to better align with the goals of BO.
Using the framework of utility-calibrated variational inference, we unify GP approximation and data acquisition into a joint optimization problem.
arXiv Detail & Related papers (2024-06-06T17:55:02Z) - Cost-Sensitive Multi-Fidelity Bayesian Optimization with Transfer of Learning Curve Extrapolation [55.75188191403343]
We introduce utility, which is a function predefined by each user and describes the trade-off between cost and performance of BO.
We validate our algorithm on various LC datasets and found it outperform all the previous multi-fidelity BO and transfer-BO baselines we consider.
arXiv Detail & Related papers (2024-05-28T07:38:39Z) - Poisson Process for Bayesian Optimization [126.51200593377739]
We propose a ranking-based surrogate model based on the Poisson process and introduce an efficient BO framework, namely Poisson Process Bayesian Optimization (PoPBO)
Compared to the classic GP-BO method, our PoPBO has lower costs and better robustness to noise, which is verified by abundant experiments.
arXiv Detail & Related papers (2024-02-05T02:54:50Z) - A General Framework for User-Guided Bayesian Optimization [51.96352579696041]
We propose ColaBO, the first Bayesian-principled framework for prior beliefs beyond the typical kernel structure.
We empirically demonstrate ColaBO's ability to substantially accelerate optimization when the prior information is accurate, and to retain approximately default performance when it is misleading.
arXiv Detail & Related papers (2023-11-24T18:27:26Z) - Towards Automated Design of Bayesian Optimization via Exploratory
Landscape Analysis [11.143778114800272]
We show that a dynamic selection of the AF can benefit the BO design.
We pave a way towards AutoML-assisted, on-the-fly BO designs that adjust their behavior on a run-by-run basis.
arXiv Detail & Related papers (2022-11-17T17:15:04Z) - Generalizing Bayesian Optimization with Decision-theoretic Entropies [102.82152945324381]
We consider a generalization of Shannon entropy from work in statistical decision theory.
We first show that special cases of this entropy lead to popular acquisition functions used in BO procedures.
We then show how alternative choices for the loss yield a flexible family of acquisition functions.
arXiv Detail & Related papers (2022-10-04T04:43:58Z) - Pre-training helps Bayesian optimization too [49.28382118032923]
We seek an alternative practice for setting functional priors.
In particular, we consider the scenario where we have data from similar functions that allow us to pre-train a tighter distribution a priori.
Our results show that our method is able to locate good hyper parameters at least 3 times more efficiently than the best competing methods.
arXiv Detail & Related papers (2022-07-07T04:42:54Z) - Bayesian Optimization over Permutation Spaces [30.650753803587794]
We propose and evaluate two algorithms for BO over Permutation Spaces (BOPS)
We theoretically analyze the performance of BOPS-T to show that their regret grows sub-linearly.
Our experiments on multiple synthetic and real-world benchmarks show that both BOPS-T and BOPS-H perform better than the state-of-the-art BO algorithm for spaces.
arXiv Detail & Related papers (2021-12-02T08:20:50Z) - One-parameter family of acquisition functions for efficient global
optimization [0.0]
We propose a new one- parameter family of acquisition functions for BO that unifies EI and PI.
The proposed method is numerically inexpensive, is easy to implement, can be easily parallelized, and on benchmark tasks shows a performance superior to EI and GP-UCB.
arXiv Detail & Related papers (2021-04-26T06:41:30Z) - Revisiting Bayesian Optimization in the light of the COCO benchmark [1.4467794332678539]
This article reports a large investigation about the effects on the performance of (Gaussian process based) BO of common and less common design choices.
The code developed for this study makes the new version (v2.1.1) of the R package DiceOptim available on CRAN.
arXiv Detail & Related papers (2021-03-30T19:45:18Z) - Incorporating Expert Prior in Bayesian Optimisation via Space Warping [54.412024556499254]
In big search spaces the algorithm goes through several low function value regions before reaching the optimum of the function.
One approach to subside this cold start phase is to use prior knowledge that can accelerate the optimisation.
In this paper, we represent the prior knowledge about the function optimum through a prior distribution.
The prior distribution is then used to warp the search space in such a way that space gets expanded around the high probability region of function optimum and shrinks around low probability region of optimum.
arXiv Detail & Related papers (2020-03-27T06:18:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.