Look-Ahead Acquisition Functions for Bernoulli Level Set Estimation
- URL: http://arxiv.org/abs/2203.09751v1
- Date: Fri, 18 Mar 2022 05:25:35 GMT
- Title: Look-Ahead Acquisition Functions for Bernoulli Level Set Estimation
- Authors: Benjamin Letham, Phillip Guan, Chase Tymms, Eytan Bakshy, Michael
Shvartsman
- Abstract summary: We derive analytic expressions for look-ahead posteriors of sublevel set membership.
We show how these lead to analytic expressions for a class of look-ahead LSE acquisition functions.
- Score: 9.764638397706717
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Level set estimation (LSE) is the problem of identifying regions where an
unknown function takes values above or below a specified threshold. Active
sampling strategies for efficient LSE have primarily been studied in
continuous-valued functions. Motivated by applications in human psychophysics
where common experimental designs produce binary responses, we study LSE active
sampling with Bernoulli outcomes. With Gaussian process classification
surrogate models, the look-ahead model posteriors used by state-of-the-art
continuous-output methods are intractable. However, we derive analytic
expressions for look-ahead posteriors of sublevel set membership, and show how
these lead to analytic expressions for a class of look-ahead LSE acquisition
functions, including information-based methods. Benchmark experiments show the
importance of considering the global look-ahead impact on the entire posterior.
We demonstrate a clear benefit to using this new class of acquisition functions
on benchmark problems, and on a challenging real-world task of estimating a
high-dimensional contrast sensitivity function.
Related papers
- $f$-Divergence Based Classification: Beyond the Use of Cross-Entropy [4.550290285002704]
We adopt a Bayesian perspective and formulate the classification task as a maximum a posteriori probability problem.
We propose a class of objective functions based on the variational representation of the $f$-divergence.
We theoretically analyze the objective functions proposed and numerically test them in three application scenarios.
arXiv Detail & Related papers (2024-01-02T16:14:02Z) - Cross-functional Analysis of Generalisation in Behavioural Learning [4.0810783261728565]
We introduce BeLUGA, an analysis method for evaluating behavioural learning considering generalisation across dimensions of different levels.
An aggregate score measures generalisation to unseen functionalities (or overfitting)
arXiv Detail & Related papers (2023-05-22T11:54:19Z) - Promises and Pitfalls of the Linearized Laplace in Bayesian Optimization [73.80101701431103]
The linearized-Laplace approximation (LLA) has been shown to be effective and efficient in constructing Bayesian neural networks.
We study the usefulness of the LLA in Bayesian optimization and highlight its strong performance and flexibility.
arXiv Detail & Related papers (2023-04-17T14:23:43Z) - Sensing Cox Processes via Posterior Sampling and Positive Bases [56.82162768921196]
We study adaptive sensing of point processes, a widely used model from spatial statistics.
We model the intensity function as a sample from a truncated Gaussian process, represented in a specially constructed positive basis.
Our adaptive sensing algorithms use Langevin dynamics and are based on posterior sampling (textscCox-Thompson) and top-two posterior sampling (textscTop2) principles.
arXiv Detail & Related papers (2021-10-21T14:47:06Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Effective Evaluation of Deep Active Learning on Image Classification
Tasks [10.27095298129151]
We present a unified re-implementation of state-of-the-art active learning algorithms in the context of image classification.
On the positive side, we show that AL techniques are 2x to 4x more label-efficient compared to RS with the use of data augmentation.
arXiv Detail & Related papers (2021-06-16T23:29:39Z) - An Analysis of State-of-the-art Activation Functions For Supervised Deep
Neural Network [11.363694566368153]
This paper provides an analysis of state-of-the-art activation functions with respect to supervised classification of deep neural network.
Experiments over two deep learning network architectures integrating these activation functions are conducted.
arXiv Detail & Related papers (2021-04-05T16:50:31Z) - An Information-Theoretic Framework for Unifying Active Learning Problems [44.758281991246825]
This paper presents an information-theoretic framework for unifying active learning problems.
We first introduce a novel active learning criterion that subsumes an existing LSE algorithm.
By exploiting the relationship between LSE and BO, we design a competitive information-theoretic acquisition function for BO.
arXiv Detail & Related papers (2020-12-19T14:22:48Z) - High Dimensional Level Set Estimation with Bayesian Neural Network [58.684954492439424]
This paper proposes novel methods to solve the high dimensional Level Set Estimation problems using Bayesian Neural Networks.
For each problem, we derive the corresponding theoretic information based acquisition function to sample the data points.
Numerical experiments on both synthetic and real-world datasets show that our proposed method can achieve better results compared to existing state-of-the-art approaches.
arXiv Detail & Related papers (2020-12-17T23:21:53Z) - Extreme Memorization via Scale of Initialization [72.78162454173803]
We construct an experimental setup in which changing the scale of initialization strongly impacts the implicit regularization induced by SGD.
We find that the extent and manner in which generalization ability is affected depends on the activation and loss function used.
In the case of the homogeneous ReLU activation, we show that this behavior can be attributed to the loss function.
arXiv Detail & Related papers (2020-08-31T04:53:11Z) - Estimating Structural Target Functions using Machine Learning and
Influence Functions [103.47897241856603]
We propose a new framework for statistical machine learning of target functions arising as identifiable functionals from statistical models.
This framework is problem- and model-agnostic and can be used to estimate a broad variety of target parameters of interest in applied statistics.
We put particular focus on so-called coarsening at random/doubly robust problems with partially unobserved information.
arXiv Detail & Related papers (2020-08-14T16:48:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.