Global Optimization of Gaussian processes
- URL: http://arxiv.org/abs/2005.10902v1
- Date: Thu, 21 May 2020 20:59:11 GMT
- Title: Global Optimization of Gaussian processes
- Authors: Artur M. Schweidtmann, Dominik Bongartz, Daniel Grothe, Tim
Kerkenhoff, Xiaopeng Lin, Jaromil Najman, Alexander Mitsos
- Abstract summary: We propose a reduced-space formulation with trained Gaussian processes trained on few data points.
The approach also leads to significantly smaller and computationally cheaper sub solver for lower bounding.
In total, we reduce time convergence by orders of orders of the proposed method.
- Score: 52.77024349608834
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Gaussian processes~(Kriging) are interpolating data-driven models that are
frequently applied in various disciplines. Often, Gaussian processes are
trained on datasets and are subsequently embedded as surrogate models in
optimization problems. These optimization problems are nonconvex and global
optimization is desired. However, previous literature observed computational
burdens limiting deterministic global optimization to Gaussian processes
trained on few data points. We propose a reduced-space formulation for
deterministic global optimization with trained Gaussian processes embedded. For
optimization, the branch-and-bound solver branches only on the degrees of
freedom and McCormick relaxations are propagated through explicit Gaussian
process models. The approach also leads to significantly smaller and
computationally cheaper subproblems for lower and upper bounding. To further
accelerate convergence, we derive envelopes of common covariance functions for
GPs and tight relaxations of acquisition functions used in Bayesian
optimization including expected improvement, probability of improvement, and
lower confidence bound. In total, we reduce computational time by orders of
magnitude compared to state-of-the-art methods, thus overcoming previous
computational burdens. We demonstrate the performance and scaling of the
proposed method and apply it to Bayesian optimization with global optimization
of the acquisition function and chance-constrained programming. The Gaussian
process models, acquisition functions, and training scripts are available
open-source within the "MeLOn - Machine Learning Models for Optimization"
toolbox~(https://git.rwth-aachen.de/avt.svt/public/MeLOn).
Related papers
- Sample-efficient Bayesian Optimisation Using Known Invariances [56.34916328814857]
We show that vanilla and constrained BO algorithms are inefficient when optimising invariant objectives.
We derive a bound on the maximum information gain of these invariant kernels.
We use our method to design a current drive system for a nuclear fusion reactor, finding a high-performance solution.
arXiv Detail & Related papers (2024-10-22T12:51:46Z) - Global Optimization of Gaussian Process Acquisition Functions Using a Piecewise-Linear Kernel Approximation [2.3342885570554652]
We introduce a piecewise approximation for process kernels and a corresponding MIQP representation for acquisition functions.
We empirically demonstrate the framework on synthetic functions, constrained benchmarks, and hyper tuning tasks.
arXiv Detail & Related papers (2024-10-22T10:56:52Z) - Enhancing Gaussian Process Surrogates for Optimization and Posterior Approximation via Random Exploration [2.984929040246293]
novel noise-free Bayesian optimization strategies that rely on a random exploration step to enhance the accuracy of Gaussian process surrogate models.
New algorithms retain the ease of implementation of the classical GP-UCB, but an additional exploration step facilitates their convergence.
arXiv Detail & Related papers (2024-01-30T14:16:06Z) - Extrinsic Bayesian Optimizations on Manifolds [1.3477333339913569]
We propose an extrinsic Bayesian optimization (eBO) framework for general optimization problems on Euclid manifold.
Our approach is to employ extrinsic Gaussian processes by first embedding the manifold onto some higher dimensionalean space.
This leads to efficient and scalable algorithms for optimization over complex manifold.
arXiv Detail & Related papers (2022-12-21T06:10:12Z) - Gaussian Processes and Statistical Decision-making in Non-Euclidean
Spaces [96.53463532832939]
We develop techniques for broadening the applicability of Gaussian processes.
We introduce a wide class of efficient approximations built from this viewpoint.
We develop a collection of Gaussian process models over non-Euclidean spaces.
arXiv Detail & Related papers (2022-02-22T01:42:57Z) - Reducing the Variance of Gaussian Process Hyperparameter Optimization
with Preconditioning [54.01682318834995]
Preconditioning is a highly effective step for any iterative method involving matrix-vector multiplication.
We prove that preconditioning has an additional benefit that has been previously unexplored.
It simultaneously can reduce variance at essentially negligible cost.
arXiv Detail & Related papers (2021-07-01T06:43:11Z) - Hyper-optimization with Gaussian Process and Differential Evolution
Algorithm [0.0]
This paper presents specific modifications of Gaussian Process optimization components from available scientific libraries.
presented modifications were submitted to BlackBox 2020 challenge, where it outperformed some conventionally available optimization libraries.
arXiv Detail & Related papers (2021-01-26T08:33:00Z) - Recent Theoretical Advances in Non-Convex Optimization [56.88981258425256]
Motivated by recent increased interest in analysis of optimization algorithms for non- optimization in deep networks and other problems in data, we give an overview of recent results of theoretical optimization algorithms for non- optimization.
arXiv Detail & Related papers (2020-12-11T08:28:51Z) - Real-Time Optimization Meets Bayesian Optimization and Derivative-Free
Optimization: A Tale of Modifier Adaptation [0.0]
This paper investigates a new class of modifier-adaptation schemes to overcome plant-model mismatch in real-time optimization of uncertain processes.
The proposed schemes embed a physical model and rely on trust-region ideas to minimize risk during the exploration.
The benefits of using an acquisition function, knowing the process noise level, or specifying a nominal process model are illustrated.
arXiv Detail & Related papers (2020-09-18T12:57:17Z) - A Primer on Zeroth-Order Optimization in Signal Processing and Machine
Learning [95.85269649177336]
ZO optimization iteratively performs three major steps: gradient estimation, descent direction, and solution update.
We demonstrate promising applications of ZO optimization, such as evaluating and generating explanations from black-box deep learning models, and efficient online sensor management.
arXiv Detail & Related papers (2020-06-11T06:50:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.