Active Learning in Bayesian Neural Networks with Balanced Entropy
Learning Principle
- URL: http://arxiv.org/abs/2105.14559v3
- Date: Sat, 15 Apr 2023 05:45:25 GMT
- Title: Active Learning in Bayesian Neural Networks with Balanced Entropy
Learning Principle
- Authors: Jae Oh Woo
- Abstract summary: We propose a new uncertainty measure, Balanced Entropy Acquisition (BalEntAcq), which captures the information balance between the uncertainty of underlying softmax probability and the label variable.
BalEntAcq consistently outperforms well-known linearly scalable active learning methods.
- Score: 0.8122270502556371
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Acquiring labeled data is challenging in many machine learning applications
with limited budgets. Active learning gives a procedure to select the most
informative data points and improve data efficiency by reducing the cost of
labeling. The info-max learning principle maximizing mutual information such as
BALD has been successful and widely adapted in various active learning
applications. However, this pool-based specific objective inherently introduces
a redundant selection and further requires a high computational cost for batch
selection. In this paper, we design and propose a new uncertainty measure,
Balanced Entropy Acquisition (BalEntAcq), which captures the information
balance between the uncertainty of underlying softmax probability and the label
variable. To do this, we approximate each marginal distribution by Beta
distribution. Beta approximation enables us to formulate BalEntAcq as a ratio
between an augmented entropy and the marginalized joint entropy. The
closed-form expression of BalEntAcq facilitates parallelization by estimating
two parameters in each marginal Beta distribution. BalEntAcq is a purely
standalone measure without requiring any relational computations with other
data points. Nevertheless, BalEntAcq captures a well-diversified selection near
the decision boundary with a margin, unlike other existing uncertainty measures
such as BALD, Entropy, or Mean Standard Deviation (MeanSD). Finally, we
demonstrate that our balanced entropy learning principle with BalEntAcq
consistently outperforms well-known linearly scalable active learning methods,
including a recently proposed PowerBALD, a simple but diversified version of
BALD, by showing experimental results obtained from MNIST, CIFAR-100, SVHN, and
TinyImageNet datasets.
Related papers
- A Bayesian Approach to Data Point Selection [24.98069363998565]
Data point selection (DPS) is becoming a critical topic in deep learning.
Existing approaches to DPS are predominantly based on a bi-level optimisation (BLO) formulation.
We propose a novel Bayesian approach to DPS.
arXiv Detail & Related papers (2024-11-06T09:04:13Z) - Equation Discovery with Bayesian Spike-and-Slab Priors and Efficient Kernels [57.46832672991433]
We propose a novel equation discovery method based on Kernel learning and BAyesian Spike-and-Slab priors (KBASS)
We use kernel regression to estimate the target function, which is flexible, expressive, and more robust to data sparsity and noises.
We develop an expectation-propagation expectation-maximization algorithm for efficient posterior inference and function estimation.
arXiv Detail & Related papers (2023-10-09T03:55:09Z) - Symmetric Equilibrium Learning of VAEs [56.56929742714685]
We view variational autoencoders (VAEs) as decoder-encoder pairs, which map distributions in the data space to distributions in the latent space and vice versa.
We propose a Nash equilibrium learning approach, which is symmetric with respect to the encoder and decoder and allows learning VAEs in situations where both the data and the latent distributions are accessible only by sampling.
arXiv Detail & Related papers (2023-07-19T10:27:34Z) - Adaptive Conformal Prediction by Reweighting Nonconformity Score [0.0]
We use a Quantile Regression Forest (QRF) to learn the distribution of nonconformity scores and utilize the QRF's weights to assign more importance to samples with residuals similar to the test point.
Our approach enjoys an assumption-free finite sample marginal and training-conditional coverage, and under suitable assumptions, it also ensures conditional coverage.
arXiv Detail & Related papers (2023-03-22T16:42:19Z) - Sample-based Uncertainty Quantification with a Single Deterministic
Neural Network [4.7210697296108926]
We propose an improved neural architecture of DISCO Nets that admits a more stable and smooth training.
We provide a new elementary proof for the validity of using the energy score to learn predictive distributions.
arXiv Detail & Related papers (2022-09-17T22:37:39Z) - Bayesian Active Learning with Fully Bayesian Gaussian Processes [0.0]
In active learning, where labeled data is scarce or difficult to obtain, neglecting this trade-off can cause inefficient querying.
We show that incorporating the bias-variance trade-off in the acquisition functions mitigates unnecessary and expensive data labeling.
arXiv Detail & Related papers (2022-05-20T13:52:04Z) - Efficient Model-based Multi-agent Reinforcement Learning via Optimistic
Equilibrium Computation [93.52573037053449]
H-MARL (Hallucinated Multi-Agent Reinforcement Learning) learns successful equilibrium policies after a few interactions with the environment.
We demonstrate our approach experimentally on an autonomous driving simulation benchmark.
arXiv Detail & Related papers (2022-03-14T17:24:03Z) - Maximizing Conditional Entropy for Batch-Mode Active Learning of
Perceptual Metrics [14.777274711706653]
We present a novel approach for batch mode active metric learning using the Maximum Entropy Principle.
We take advantage of the monotonically increasing submodular entropy function to construct an efficient greedy algorithm.
Our approach is the first batch-mode active metric learning method to define a unified score that balances informativeness and diversity for an entire batch of triplets.
arXiv Detail & Related papers (2021-02-15T06:55:17Z) - Attentional-Biased Stochastic Gradient Descent [74.49926199036481]
We present a provable method (named ABSGD) for addressing the data imbalance or label noise problem in deep learning.
Our method is a simple modification to momentum SGD where we assign an individual importance weight to each sample in the mini-batch.
ABSGD is flexible enough to combine with other robust losses without any additional cost.
arXiv Detail & Related papers (2020-12-13T03:41:52Z) - Sparse Feature Selection Makes Batch Reinforcement Learning More Sample
Efficient [62.24615324523435]
This paper provides a statistical analysis of high-dimensional batch Reinforcement Learning (RL) using sparse linear function approximation.
When there is a large number of candidate features, our result sheds light on the fact that sparsity-aware methods can make batch RL more sample efficient.
arXiv Detail & Related papers (2020-11-08T16:48:02Z) - Probabilistic Case-based Reasoning for Open-World Knowledge Graph
Completion [59.549664231655726]
A case-based reasoning (CBR) system solves a new problem by retrieving cases' that are similar to the given problem.
In this paper, we demonstrate that such a system is achievable for reasoning in knowledge-bases (KBs)
Our approach predicts attributes for an entity by gathering reasoning paths from similar entities in the KB.
arXiv Detail & Related papers (2020-10-07T17:48:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.