Influence Selection for Active Learning
- URL: http://arxiv.org/abs/2108.09331v1
- Date: Fri, 20 Aug 2021 18:44:52 GMT
- Title: Influence Selection for Active Learning
- Authors: Zhuoming Liu, Hao Ding, Huaping Zhong, Weijia Li, Jifeng Dai, Conghui
He
- Abstract summary: We propose the Influence Selection for Active Learning(ISAL) which selects the unlabeled samples that can provide the most positive Influence on model performance.
ISAL achieves state-of-the-art performance in different active learning settings for different tasks with different datasets.
- Score: 18.84939869188145
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The existing active learning methods select the samples by evaluating the
sample's uncertainty or its effect on the diversity of labeled datasets based
on different task-specific or model-specific criteria. In this paper, we
propose the Influence Selection for Active Learning(ISAL) which selects the
unlabeled samples that can provide the most positive Influence on model
performance. To obtain the Influence of the unlabeled sample in the active
learning scenario, we design the Untrained Unlabeled sample Influence
Calculation(UUIC) to estimate the unlabeled sample's expected gradient with
which we calculate its Influence. To prove the effectiveness of UUIC, we
provide both theoretical and experimental analyses. Since the UUIC just depends
on the model gradients, which can be obtained easily from any neural network,
our active learning algorithm is task-agnostic and model-agnostic. ISAL
achieves state-of-the-art performance in different active learning settings for
different tasks with different datasets. Compared with previous methods, our
method decreases the annotation cost at least by 12%, 13% and 16% on CIFAR10,
VOC2012 and COCO, respectively.
Related papers
- Understanding Uncertainty-based Active Learning Under Model Mismatch [16.361254095103615]
Uncertainty-based Active Learning (UAL) operates by querying the label(s) of pivotal samples from an unlabeled pool selected based on the prediction uncertainty.
The efficacy of UAL depends on the model capacity as well as the adopted uncertainty-based acquisition function.
arXiv Detail & Related papers (2024-08-24T23:37:08Z) - Provably Neural Active Learning Succeeds via Prioritizing Perplexing Samples [53.95282502030541]
Neural Network-based active learning (NAL) is a cost-effective data selection technique that utilizes neural networks to select and train on a small subset of samples.
We try to move one step forward by offering a unified explanation for the success of both query criteria-based NAL from a feature learning view.
arXiv Detail & Related papers (2024-06-06T10:38:01Z) - Querying Easily Flip-flopped Samples for Deep Active Learning [63.62397322172216]
Active learning is a machine learning paradigm that aims to improve the performance of a model by strategically selecting and querying unlabeled data.
One effective selection strategy is to base it on the model's predictive uncertainty, which can be interpreted as a measure of how informative a sample is.
This paper proposes the it least disagree metric (LDM) as the smallest probability of disagreement of the predicted label.
arXiv Detail & Related papers (2024-01-18T08:12:23Z) - BAL: Balancing Diversity and Novelty for Active Learning [53.289700543331925]
We introduce a novel framework, Balancing Active Learning (BAL), which constructs adaptive sub-pools to balance diverse and uncertain data.
Our approach outperforms all established active learning methods on widely recognized benchmarks by 1.20%.
arXiv Detail & Related papers (2023-12-26T08:14:46Z) - Learning Objective-Specific Active Learning Strategies with Attentive
Neural Processes [72.75421975804132]
Learning Active Learning (LAL) suggests to learn the active learning strategy itself, allowing it to adapt to the given setting.
We propose a novel LAL method for classification that exploits symmetry and independence properties of the active learning problem.
Our approach is based on learning from a myopic oracle, which gives our model the ability to adapt to non-standard objectives.
arXiv Detail & Related papers (2023-09-11T14:16:37Z) - Optimal Sample Selection Through Uncertainty Estimation and Its
Application in Deep Learning [22.410220040736235]
We present a theoretically optimal solution for addressing both coreset selection and active learning.
Our proposed method, COPS, is designed to minimize the expected loss of a model trained on subsampled data.
arXiv Detail & Related papers (2023-09-05T14:06:33Z) - Temporal Output Discrepancy for Loss Estimation-based Active Learning [65.93767110342502]
We present a novel deep active learning approach that queries the oracle for data annotation when the unlabeled sample is believed to incorporate high loss.
Our approach achieves superior performances than the state-of-the-art active learning methods on image classification and semantic segmentation tasks.
arXiv Detail & Related papers (2022-12-20T19:29:37Z) - Improving Robustness and Efficiency in Active Learning with Contrastive
Loss [13.994967246046008]
This paper introduces supervised contrastive active learning (SCAL) by leveraging the contrastive loss for active learning in a supervised setting.
We propose efficient query strategies in active learning to select unbiased and informative data samples of diverse feature representations.
arXiv Detail & Related papers (2021-09-13T21:09:21Z) - Minimax Active Learning [61.729667575374606]
Active learning aims to develop label-efficient algorithms by querying the most representative samples to be labeled by a human annotator.
Current active learning techniques either rely on model uncertainty to select the most uncertain samples or use clustering or reconstruction to choose the most diverse set of unlabeled examples.
We develop a semi-supervised minimax entropy-based active learning algorithm that leverages both uncertainty and diversity in an adversarial manner.
arXiv Detail & Related papers (2020-12-18T19:03:40Z) - Deep Active Learning for Sequence Labeling Based on Diversity and
Uncertainty in Gradient [5.33024001730262]
We show that the amount of labeled training data can be reduced using active learning when it incorporates both uncertainty and diversity in the sequence labeling task.
We examined the effects of our sequence-based approach by selecting weighted diverse in the gradient embedding approach across multiple tasks, datasets, models, and consistently outperform classic uncertainty-based sampling and diversity-based sampling.
arXiv Detail & Related papers (2020-11-27T06:03:27Z) - Ask-n-Learn: Active Learning via Reliable Gradient Representations for
Image Classification [29.43017692274488]
Deep predictive models rely on human supervision in the form of labeled training data.
We propose Ask-n-Learn, an active learning approach based on gradient embeddings obtained using the pesudo-labels estimated in each of the algorithm.
arXiv Detail & Related papers (2020-09-30T05:19:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.