Stochastic Batch Acquisition: A Simple Baseline for Deep Active Learning
- URL: http://arxiv.org/abs/2106.12059v3
- Date: Tue, 19 Sep 2023 21:20:38 GMT
- Title: Stochastic Batch Acquisition: A Simple Baseline for Deep Active Learning
- Authors: Andreas Kirsch, Sebastian Farquhar, Parmida Atighehchian, Andrew
Jesson, Frederic Branchaud-Charron, Yarin Gal
- Abstract summary: We examine a simple strategy for adapting well-known single-point acquisition functions to allow batch active learning.
This strategy can perform just as well as compute-intensive state-of-the-art batch acquisition functions, like BatchBALD or BADGE, while using orders of magnitude less compute.
In addition to providing a practical option for machine learning practitioners, the surprising success of the proposed method in a wide range of experimental settings raises a difficult question for the field.
- Score: 48.19646855997791
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We examine a simple stochastic strategy for adapting well-known single-point
acquisition functions to allow batch active learning. Unlike acquiring the
top-K points from the pool set, score- or rank-based sampling takes into
account that acquisition scores change as new data are acquired. This simple
strategy for adapting standard single-sample acquisition strategies can even
perform just as well as compute-intensive state-of-the-art batch acquisition
functions, like BatchBALD or BADGE, while using orders of magnitude less
compute. In addition to providing a practical option for machine learning
practitioners, the surprising success of the proposed method in a wide range of
experimental settings raises a difficult question for the field: when are these
expensive batch acquisition methods pulling their weight?
Related papers
- Novel Batch Active Learning Approach and Its Application to Synthetic
Aperture Radar Datasets [7.381841249558068]
Recent gains have been made using sequential active learning for synthetic aperture radar (SAR) data arXiv:2204.00005.
We developed a novel, two-part approach for batch active learning: Dijkstra's Annulus Core-Set (DAC) for core-set generation and LocalMax for batch sampling.
The batch active learning process that combines DAC and LocalMax achieves nearly identical accuracy as sequential active learning but is more efficient, proportional to the batch size.
arXiv Detail & Related papers (2023-07-19T23:25:21Z) - BatchGFN: Generative Flow Networks for Batch Active Learning [80.73649229919454]
BatchGFN is a novel approach for pool-based active learning that uses generative flow networks to sample sets of data points proportional to a batch reward.
We show our approach enables principled sampling near-optimal utility batches at inference time with a single forward pass per point in the batch in toy regression problems.
arXiv Detail & Related papers (2023-06-26T20:41:36Z) - An Embarrassingly Simple Approach to Semi-Supervised Few-Shot Learning [58.59343434538218]
We propose a simple but quite effective approach to predict accurate negative pseudo-labels of unlabeled data from an indirect learning perspective.
Our approach can be implemented in just few lines of code by only using off-the-shelf operations.
arXiv Detail & Related papers (2022-09-28T02:11:34Z) - Achieving Minimax Rates in Pool-Based Batch Active Learning [26.12124106759262]
We consider a batch active learning scenario where the learner adaptively issues batches of points to a labeling oracle.
In this paper we propose a solution which requires a careful trade off between the informativeness of the queried points and their diversity.
arXiv Detail & Related papers (2022-02-11T04:55:45Z) - Batch Active Learning at Scale [39.26441165274027]
Batch active learning, which adaptively issues batched queries to a labeling oracle, is a common approach for addressing this problem.
In this work, we analyze an efficient active learning algorithm, which focuses on the large batch setting.
We show that our sampling method, which combines notions of uncertainty and diversity, easily scales to batch sizes (100K-1M) several orders of magnitude larger than used in previous studies.
arXiv Detail & Related papers (2021-07-29T18:14:05Z) - Few-shot Sequence Learning with Transformers [79.87875859408955]
Few-shot algorithms aim at learning new tasks provided only a handful of training examples.
In this work we investigate few-shot learning in the setting where the data points are sequences of tokens.
We propose an efficient learning algorithm based on Transformers.
arXiv Detail & Related papers (2020-12-17T12:30:38Z) - Semi-supervised Batch Active Learning via Bilevel Optimization [89.37476066973336]
We formulate our approach as a data summarization problem via bilevel optimization.
We show that our method is highly effective in keyword detection tasks in the regime when only few labeled samples are available.
arXiv Detail & Related papers (2020-10-19T16:53:24Z) - Zeroth-Order Supervised Policy Improvement [94.0748002906652]
Policy gradient (PG) algorithms have been widely used in reinforcement learning (RL)
We propose Zeroth-Order Supervised Policy Improvement (ZOSPI)
ZOSPI exploits the estimated value function $Q$ globally while preserving the local exploitation of the PG methods.
arXiv Detail & Related papers (2020-06-11T16:49:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.