On Bayesian Search for the Feasible Space Under Computationally
Expensive Constraints
- URL: http://arxiv.org/abs/2004.11055v2
- Date: Wed, 24 Jun 2020 12:00:05 GMT
- Title: On Bayesian Search for the Feasible Space Under Computationally
Expensive Constraints
- Authors: Alma Rahat and Michael Wood
- Abstract summary: We propose a novel acquisition function that combines the probability that a solution lies at the boundary between feasible and infeasible spaces.
Experiments confirmed the efficacy of the proposed function.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We are often interested in identifying the feasible subset of a decision
space under multiple constraints to permit effective design exploration. If
determining feasibility required computationally expensive simulations, the
cost of exploration would be prohibitive. Bayesian search is data-efficient for
such problems: starting from a small dataset, the central concept is to use
Bayesian models of constraints with an acquisition function to locate promising
solutions that may improve predictions of feasibility when the dataset is
augmented. At the end of this sequential active learning approach with a
limited number of expensive evaluations, the models can accurately predict the
feasibility of any solution obviating the need for full simulations. In this
paper, we propose a novel acquisition function that combines the probability
that a solution lies at the boundary between feasible and infeasible spaces
(representing exploitation) and the entropy in predictions (representing
exploration). Experiments confirmed the efficacy of the proposed function.
Related papers
- Minimax and Communication-Efficient Distributed Best Subset Selection with Oracle Property [0.358439716487063]
The explosion of large-scale data has outstripped the processing capabilities of single-machine systems.
Traditional approaches to distributed inference often struggle with achieving true sparsity in high-dimensional datasets.
We propose a novel, two-stage, distributed best subset selection algorithm to address these issues.
arXiv Detail & Related papers (2024-08-30T13:22:08Z) - Source-Free Unsupervised Domain Adaptation with Hypothesis Consolidation
of Prediction Rationale [53.152460508207184]
Source-Free Unsupervised Domain Adaptation (SFUDA) is a challenging task where a model needs to be adapted to a new domain without access to target domain labels or source domain data.
This paper proposes a novel approach that considers multiple prediction hypotheses for each sample and investigates the rationale behind each hypothesis.
To achieve the optimal performance, we propose a three-step adaptation process: model pre-adaptation, hypothesis consolidation, and semi-supervised learning.
arXiv Detail & Related papers (2024-02-02T05:53:22Z) - Calibrating Neural Simulation-Based Inference with Differentiable
Coverage Probability [50.44439018155837]
We propose to include a calibration term directly into the training objective of the neural model.
By introducing a relaxation of the classical formulation of calibration error we enable end-to-end backpropagation.
It is directly applicable to existing computational pipelines allowing reliable black-box posterior inference.
arXiv Detail & Related papers (2023-10-20T10:20:45Z) - Equation Discovery with Bayesian Spike-and-Slab Priors and Efficient Kernels [57.46832672991433]
We propose a novel equation discovery method based on Kernel learning and BAyesian Spike-and-Slab priors (KBASS)
We use kernel regression to estimate the target function, which is flexible, expressive, and more robust to data sparsity and noises.
We develop an expectation-propagation expectation-maximization algorithm for efficient posterior inference and function estimation.
arXiv Detail & Related papers (2023-10-09T03:55:09Z) - Efficient Model-Free Exploration in Low-Rank MDPs [76.87340323826945]
Low-Rank Markov Decision Processes offer a simple, yet expressive framework for RL with function approximation.
Existing algorithms are either (1) computationally intractable, or (2) reliant upon restrictive statistical assumptions.
We propose the first provably sample-efficient algorithm for exploration in Low-Rank MDPs.
arXiv Detail & Related papers (2023-07-08T15:41:48Z) - STEERING: Stein Information Directed Exploration for Model-Based
Reinforcement Learning [111.75423966239092]
We propose an exploration incentive in terms of the integral probability metric (IPM) between a current estimate of the transition model and the unknown optimal.
Based on KSD, we develop a novel algorithm algo: textbfSTEin information dirtextbfEcted exploration for model-based textbfReinforcement LearntextbfING.
arXiv Detail & Related papers (2023-01-28T00:49:28Z) - Bayesian sequential design of computer experiments for quantile set inversion [0.0]
We consider an unknown multivariate function representing a system-such as a complex numerical simulator.
Our objective is to estimate the set of deterministic inputs leading to outputs whose probability is less than a given threshold.
arXiv Detail & Related papers (2022-11-02T10:14:05Z) - Predicting the utility of search spaces for black-box optimization: a
simple, budget-aware approach [25.07599332807319]
Black box optimization requires specifying a search space to explore for solutions.
Finding a high quality search space can be challenging in many applications.
We present a simple scoring method based on a utility function applied to a probabilistic response surface model.
arXiv Detail & Related papers (2021-12-15T16:28:28Z) - Approximate Bayesian Optimisation for Neural Networks [6.921210544516486]
A body of work has been done to automate machine learning algorithm to highlight the importance of model choice.
The necessity to solve the analytical tractability and the computational feasibility in a idealistic fashion enables to ensure the efficiency and the applicability.
arXiv Detail & Related papers (2021-08-27T19:03:32Z) - Offline Model-Based Optimization via Normalized Maximum Likelihood
Estimation [101.22379613810881]
We consider data-driven optimization problems where one must maximize a function given only queries at a fixed set of points.
This problem setting emerges in many domains where function evaluation is a complex and expensive process.
We propose a tractable approximation that allows us to scale our method to high-capacity neural network models.
arXiv Detail & Related papers (2021-02-16T06:04:27Z) - Necessary and Sufficient Conditions for Inverse Reinforcement Learning
of Bayesian Stopping Time Problems [22.498689292081156]
This paper presents an inverse reinforcement learning(IRL) framework for Bayesian stopping time problems.
By observing the actions of a Bayesian decision maker, we provide a necessary and sufficient condition to identify if these actions are consistent with optimizing a cost function.
Our IRL algorithm identifies optimality and then constructs set valued estimates of the cost function.
arXiv Detail & Related papers (2020-07-07T14:14:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.