Fast Bayesian Optimization of Needle-in-a-Haystack Problems using
Zooming Memory-Based Initialization
- URL: http://arxiv.org/abs/2208.13771v1
- Date: Fri, 26 Aug 2022 23:57:41 GMT
- Title: Fast Bayesian Optimization of Needle-in-a-Haystack Problems using
Zooming Memory-Based Initialization
- Authors: Alexander E. Siemenn, Zekun Ren, Qianxiao Li, Tonio Buonassisi
- Abstract summary: A Needle-in-a-Haystack problem arises when there is an extreme imbalance of optimum conditions relative to the size of the dataset.
We present a Zooming Memory-Based Initialization algorithm that builds on conventional Bayesian optimization principles.
- Score: 73.96101108943986
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Needle-in-a-Haystack problems exist across a wide range of applications
including rare disease prediction, ecological resource management, fraud
detection, and material property optimization. A Needle-in-a-Haystack problem
arises when there is an extreme imbalance of optimum conditions relative to the
size of the dataset. For example, only 0.82% out of 146k total materials in the
open-access Materials Project database have a negative Poisson's ratio.
However, current state-of-the-art optimization algorithms are not designed with
the capabilities to find solutions to these challenging multidimensional
Needle-in-a-Haystack problems, resulting in slow convergence to a global
optimum or pigeonholing into a local minimum. In this paper, we present a
Zooming Memory-Based Initialization algorithm, entitled ZoMBI, that builds on
conventional Bayesian optimization principles to quickly and efficiently
optimize Needle-in-a-Haystack problems in both less time and fewer experiments
by addressing the common convergence and pigeonholing issues. ZoMBI actively
extracts knowledge from the previously best-performing evaluated experiments to
iteratively zoom in the sampling search bounds towards the global optimum
"needle" and then prunes the memory of low-performing historical experiments to
accelerate compute times. We validate the algorithm's performance on two
real-world 5-dimensional Needle-in-a-Haystack material property optimization
datasets: discovery of auxetic Poisson's ratio materials and discovery of high
thermoelectric figure of merit materials. The ZoMBI algorithm demonstrates
compute time speed-ups of 400x compared to traditional Bayesian optimization as
well as efficiently discovering materials in under 100 experiments that are up
to 3x more highly optimized than those discovered by current state-of-the-art
algorithms.
Related papers
- Quality-Diversity Algorithms Can Provably Be Helpful for Optimization [24.694984679399315]
Quality-Diversity (QD) algorithms aim to find a set of high-performing, yet diverse solutions.
This paper tries to shed some light on the optimization ability of QD algorithms via rigorous running time analysis.
arXiv Detail & Related papers (2024-01-19T07:40:24Z) - Characterization of Locality in Spin States and Forced Moves for
Optimizations [0.36868085124383626]
In optimization problems, the existence of local minima in energy landscapes is problematic to use to seek the global minimum.
We develop an algorithm to get out of the local minima efficiently while it does not yield the exact samplings.
As the proposed algorithm is based on a rejection-free algorithm, the computational cost is low.
arXiv Detail & Related papers (2023-12-05T07:21:00Z) - Parallel Bayesian Optimization Using Satisficing Thompson Sampling for
Time-Sensitive Black-Box Optimization [0.0]
We propose satisficing Thompson sampling-based parallel BO approaches, including synchronous and asynchronous versions.
We shift the target from an optimal solution to a satisficing solution that is easier to learn.
The effectiveness of the proposed methods is demonstrated on a fast-charging design problem of Lithium-ion batteries.
arXiv Detail & Related papers (2023-10-19T07:03:51Z) - Portfolio optimization with discrete simulated annealing [0.0]
We present an integer optimization method to find optimal portfolios in the presence of discretized convex and non- convex cost functions.
This allows us to achieve a solution with a given quality.
arXiv Detail & Related papers (2022-10-03T10:39:05Z) - Versatile Single-Loop Method for Gradient Estimator: First and Second
Order Optimality, and its Application to Federated Learning [45.78238792836363]
We present a single-loop algorithm named SLEDGE (Single-Loop-E Gradient Estimator) for periodic convergence.
Unlike existing methods, SLEDGE has the advantage of versatility; (ii) second-order optimal, (ii) in the PL region, and (iii) smaller complexity under less of data.
arXiv Detail & Related papers (2022-09-01T11:05:26Z) - Learning to Hash Robustly, with Guarantees [79.68057056103014]
In this paper, we design an NNS algorithm for the Hamming space that has worst-case guarantees essentially matching that of theoretical algorithms.
We evaluate the algorithm's ability to optimize for a given dataset both theoretically and practically.
Our algorithm has a 1.8x and 2.1x better recall on the worst-performing queries to the MNIST and ImageNet datasets.
arXiv Detail & Related papers (2021-08-11T20:21:30Z) - Bayesian Algorithm Execution: Estimating Computable Properties of
Black-box Functions Using Mutual Information [78.78486761923855]
In many real world problems, we want to infer some property of an expensive black-box function f, given a budget of T function evaluations.
We present a procedure, InfoBAX, that sequentially chooses queries that maximize mutual information with respect to the algorithm's output.
On these problems, InfoBAX uses up to 500 times fewer queries to f than required by the original algorithm.
arXiv Detail & Related papers (2021-04-19T17:22:11Z) - Towards Optimally Efficient Tree Search with Deep Learning [76.64632985696237]
This paper investigates the classical integer least-squares problem which estimates signals integer from linear models.
The problem is NP-hard and often arises in diverse applications such as signal processing, bioinformatics, communications and machine learning.
We propose a general hyper-accelerated tree search (HATS) algorithm by employing a deep neural network to estimate the optimal estimation for the underlying simplified memory-bounded A* algorithm.
arXiv Detail & Related papers (2021-01-07T08:00:02Z) - Fast Objective & Duality Gap Convergence for Non-Convex Strongly-Concave
Min-Max Problems with PL Condition [52.08417569774822]
This paper focuses on methods for solving smooth non-concave min-max problems, which have received increasing attention due to deep learning (e.g., deep AUC)
arXiv Detail & Related papers (2020-06-12T00:32:21Z) - Global Optimization of Gaussian processes [52.77024349608834]
We propose a reduced-space formulation with trained Gaussian processes trained on few data points.
The approach also leads to significantly smaller and computationally cheaper sub solver for lower bounding.
In total, we reduce time convergence by orders of orders of the proposed method.
arXiv Detail & Related papers (2020-05-21T20:59:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.