Differentiable Expected Hypervolume Improvement for Parallel
Multi-Objective Bayesian Optimization
- URL: http://arxiv.org/abs/2006.05078v3
- Date: Fri, 23 Oct 2020 04:20:57 GMT
- Title: Differentiable Expected Hypervolume Improvement for Parallel
Multi-Objective Bayesian Optimization
- Authors: Samuel Daulton, Maximilian Balandat, Eytan Bakshy
- Abstract summary: We leverage recent advances in programming models and hardware acceleration for multi-objective BO using Expected Hyper Improvement (EHVI)
We derive a novel formulation of q-Expected Hyper Improvement (qEHVI), an acquisition function that extends EHVI to the parallel, constrained evaluation setting.
Our empirical evaluation demonstrates that qEHVI is computationally tractable in many practical scenarios and outperforms state-of-the-art multi-objective BO algorithms at a fraction of their wall time.
- Score: 11.956059322407437
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In many real-world scenarios, decision makers seek to efficiently optimize
multiple competing objectives in a sample-efficient fashion. Multi-objective
Bayesian optimization (BO) is a common approach, but many of the
best-performing acquisition functions do not have known analytic gradients and
suffer from high computational overhead. We leverage recent advances in
programming models and hardware acceleration for multi-objective BO using
Expected Hypervolume Improvement (EHVI)---an algorithm notorious for its high
computational complexity. We derive a novel formulation of q-Expected
Hypervolume Improvement (qEHVI), an acquisition function that extends EHVI to
the parallel, constrained evaluation setting. qEHVI is an exact computation of
the joint EHVI of q new candidate points (up to Monte-Carlo (MC) integration
error). Whereas previous EHVI formulations rely on gradient-free acquisition
optimization or approximated gradients, we compute exact gradients of the MC
estimator via auto-differentiation, thereby enabling efficient and effective
optimization using first-order and quasi-second-order methods. Our empirical
evaluation demonstrates that qEHVI is computationally tractable in many
practical scenarios and outperforms state-of-the-art multi-objective BO
algorithms at a fraction of their wall time.
Related papers
- Unexpected Improvements to Expected Improvement for Bayesian
Optimization [23.207497480389208]
We propose LogEI, a new family of acquisition functions whose members either have identical or approximately equal optima as their canonical counterparts, but are substantially easier to optimize numerically.
Our empirical results show that members of the LogEI family of acquisition functions substantially improve on the optimization performance of their canonical counterparts and surprisingly, are on par with or exceed the performance of recent state-of-the-art acquisition functions.
arXiv Detail & Related papers (2023-10-31T17:59:56Z) - Federated Conditional Stochastic Optimization [110.513884892319]
Conditional optimization has found in a wide range of machine learning tasks, such as in-variant learning tasks, AUPRC, andAML.
This paper proposes algorithms for distributed federated learning.
arXiv Detail & Related papers (2023-10-04T01:47:37Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - Multi-objective hyperparameter optimization with performance uncertainty [62.997667081978825]
This paper presents results on multi-objective hyperparameter optimization with uncertainty on the evaluation of Machine Learning algorithms.
We combine the sampling strategy of Tree-structured Parzen Estimators (TPE) with the metamodel obtained after training a Gaussian Process Regression (GPR) with heterogeneous noise.
Experimental results on three analytical test functions and three ML problems show the improvement over multi-objective TPE and GPR.
arXiv Detail & Related papers (2022-09-09T14:58:43Z) - An Online Prediction Approach Based on Incremental Support Vector
Machine for Dynamic Multiobjective Optimization [19.336520152294213]
We propose a novel prediction algorithm based on incremental support vector machine (ISVM)
We treat the solving of dynamic multiobjective optimization problems (DMOPs) as an online learning process.
The proposed algorithm can effectively tackle dynamic multiobjective optimization problems.
arXiv Detail & Related papers (2021-02-24T08:51:23Z) - Zeroth-Order Hybrid Gradient Descent: Towards A Principled Black-Box
Optimization Framework [100.36569795440889]
This work is on the iteration of zero-th-order (ZO) optimization which does not require first-order information.
We show that with a graceful design in coordinate importance sampling, the proposed ZO optimization method is efficient both in terms of complexity as well as as function query cost.
arXiv Detail & Related papers (2020-12-21T17:29:58Z) - Batch Sequential Adaptive Designs for Global Optimization [5.825138898746968]
Efficient global optimization (EGO) is one of the most popular SAD methods for expensive black-box optimization problems.
For those multiple points EGO methods, the heavy computation and points clustering are the obstacles.
In this work, a novel batch SAD method, named "accelerated EGO", is forwarded by using a refined sampling/importance resampling (SIR) method.
The efficiency of the proposed SAD is validated by nine classic test functions with dimension from 2 to 12.
arXiv Detail & Related papers (2020-10-21T01:11:35Z) - Bilevel Optimization: Convergence Analysis and Enhanced Design [63.64636047748605]
Bilevel optimization is a tool for many machine learning problems.
We propose a novel stoc-efficientgradient estimator named stoc-BiO.
arXiv Detail & Related papers (2020-10-15T18:09:48Z) - Efficient Learning of Generative Models via Finite-Difference Score
Matching [111.55998083406134]
We present a generic strategy to efficiently approximate any-order directional derivative with finite difference.
Our approximation only involves function evaluations, which can be executed in parallel, and no gradient computations.
arXiv Detail & Related papers (2020-07-07T10:05:01Z) - Parallel Predictive Entropy Search for Multi-objective Bayesian
Optimization with Constraints [0.0]
Real-world problems often involve the optimization of several objectives under multiple constraints.
This article introduces PPESMOC, an information-based batch method for the simultaneous optimization of black-box functions.
Iteratively, PPESMOC selects a batch of input locations at which to evaluate the black-boxes.
arXiv Detail & Related papers (2020-04-01T17:37:58Z) - aphBO-2GP-3B: A budgeted asynchronous parallel multi-acquisition
functions for constrained Bayesian optimization on high-performing computing
architecture [4.738678765150249]
An asynchronous constrained batch-parallel Bayesian optimization method is proposed to solve the computationally-expensive simulation-based optimization problems.
The advantages of this method are three-fold.
The aphBO-2GP-3B framework is demonstrated using two high-fidelity expensive industrial applications.
arXiv Detail & Related papers (2020-03-20T18:02:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.