PowerEvaluationBALD: Efficient Evaluation-Oriented Deep (Bayesian)
Active Learning with Stochastic Acquisition Functions
- URL: http://arxiv.org/abs/2101.03552v1
- Date: Sun, 10 Jan 2021 13:46:45 GMT
- Title: PowerEvaluationBALD: Efficient Evaluation-Oriented Deep (Bayesian)
Active Learning with Stochastic Acquisition Functions
- Authors: Andreas Kirsch
- Abstract summary: We develop BatchEvaluationBALD, a new acquisition function for deep active learning.
We also develop a variant for the non-Bayesian setting, which we call Evaluation Information Gain.
To reduce computational requirements and allow these methods to scale to larger batch sizes, we introduce acquisition functions that use importance-sampling of tempered acquisition scores.
- Score: 2.0305676256390934
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We develop BatchEvaluationBALD, a new acquisition function for deep Bayesian
active learning, as an expansion of BatchBALD that takes into account an
evaluation set of unlabeled data, for example, the pool set. We also develop a
variant for the non-Bayesian setting, which we call Evaluation Information
Gain. To reduce computational requirements and allow these methods to scale to
larger acquisition batch sizes, we introduce stochastic acquisition functions
that use importance-sampling of tempered acquisition scores. We call this
method PowerEvaluationBALD. We show in first experiments that
PowerEvaluationBALD works on par with BatchEvaluationBALD, which outperforms
BatchBALD on Repeated MNIST (MNISTx2), while massively reducing the
computational requirements compared to BatchBALD or BatchEvaluationBALD.
Related papers
- Practical Bayesian Algorithm Execution via Posterior Sampling [24.795916177776856]
PS-BAX is a simple, effective, and scalable BAX method based on posterior sampling.
PS-BAX is applicable to a wide range of problems, including many optimization variants and level set estimation.
arXiv Detail & Related papers (2024-10-27T21:11:55Z) - Monte Carlo Tree Search Boosts Reasoning via Iterative Preference Learning [55.96599486604344]
We introduce an approach aimed at enhancing the reasoning capabilities of Large Language Models (LLMs) through an iterative preference learning process.
We use Monte Carlo Tree Search (MCTS) to iteratively collect preference data, utilizing its look-ahead ability to break down instance-level rewards into more granular step-level signals.
The proposed algorithm employs Direct Preference Optimization (DPO) to update the LLM policy using this newly generated step-level preference data.
arXiv Detail & Related papers (2024-05-01T11:10:24Z) - Poisson Process for Bayesian Optimization [126.51200593377739]
We propose a ranking-based surrogate model based on the Poisson process and introduce an efficient BO framework, namely Poisson Process Bayesian Optimization (PoPBO)
Compared to the classic GP-BO method, our PoPBO has lower costs and better robustness to noise, which is verified by abundant experiments.
arXiv Detail & Related papers (2024-02-05T02:54:50Z) - Value-Distributional Model-Based Reinforcement Learning [59.758009422067]
Quantifying uncertainty about a policy's long-term performance is important to solve sequential decision-making tasks.
We study the problem from a model-based Bayesian reinforcement learning perspective.
We propose Epistemic Quantile-Regression (EQR), a model-based algorithm that learns a value distribution function.
arXiv Detail & Related papers (2023-08-12T14:59:19Z) - BatchGFN: Generative Flow Networks for Batch Active Learning [80.73649229919454]
BatchGFN is a novel approach for pool-based active learning that uses generative flow networks to sample sets of data points proportional to a batch reward.
We show our approach enables principled sampling near-optimal utility batches at inference time with a single forward pass per point in the batch in toy regression problems.
arXiv Detail & Related papers (2023-06-26T20:41:36Z) - Diversity Enhanced Active Learning with Strictly Proper Scoring Rules [4.81450893955064]
We study acquisition functions for active learning (AL) for text classification.
We convert the Expected Loss Reduction (ELR) method to estimate the increase in (strictly proper) scores like log probability or negative mean square error.
We show that the use of mean square error and log probability with BEMPS yields robust acquisition functions.
arXiv Detail & Related papers (2021-10-27T05:02:11Z) - Estimation Error Correction in Deep Reinforcement Learning for
Deterministic Actor-Critic Methods [0.0]
In value-based deep reinforcement learning methods, approximation of value functions induces overestimation bias and leads to suboptimal policies.
We show that in deep actor-critic methods that aim to overcome the overestimation bias, if the reinforcement signals received by the agent have a high variance, a significant underestimation bias arises.
To minimize the underestimation, we introduce a parameter-free, novel deep Q-learning variant.
arXiv Detail & Related papers (2021-09-22T13:49:35Z) - Stochastic Batch Acquisition: A Simple Baseline for Deep Active Learning [48.19646855997791]
We examine a simple strategy for adapting well-known single-point acquisition functions to allow batch active learning.
This strategy can perform just as well as compute-intensive state-of-the-art batch acquisition functions, like BatchBALD or BADGE, while using orders of magnitude less compute.
In addition to providing a practical option for machine learning practitioners, the surprising success of the proposed method in a wide range of experimental settings raises a difficult question for the field.
arXiv Detail & Related papers (2021-06-22T21:07:50Z) - Semi-supervised Batch Active Learning via Bilevel Optimization [89.37476066973336]
We formulate our approach as a data summarization problem via bilevel optimization.
We show that our method is highly effective in keyword detection tasks in the regime when only few labeled samples are available.
arXiv Detail & Related papers (2020-10-19T16:53:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.