When Deep Learners Change Their Mind: Learning Dynamics for Active
Learning
- URL: http://arxiv.org/abs/2107.14707v1
- Date: Fri, 30 Jul 2021 15:30:17 GMT
- Title: When Deep Learners Change Their Mind: Learning Dynamics for Active
Learning
- Authors: Javad Zolfaghari Bengar, Bogdan Raducanu, Joost van de Weijer
- Abstract summary: In this paper, we propose a new informativeness-based active learning method.
Our measure is derived from the learning dynamics of a neural network.
We show that label-dispersion is a promising predictor of the uncertainty of the network.
- Score: 32.792098711779424
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Active learning aims to select samples to be annotated that yield the largest
performance improvement for the learning algorithm. Many methods approach this
problem by measuring the informativeness of samples and do this based on the
certainty of the network predictions for samples. However, it is well-known
that neural networks are overly confident about their prediction and are
therefore an untrustworthy source to assess sample informativeness. In this
paper, we propose a new informativeness-based active learning method. Our
measure is derived from the learning dynamics of a neural network. More
precisely we track the label assignment of the unlabeled data pool during the
training of the algorithm. We capture the learning dynamics with a metric
called label-dispersion, which is low when the network consistently assigns the
same label to the sample during the training of the network and high when the
assigned label changes frequently. We show that label-dispersion is a promising
predictor of the uncertainty of the network, and show on two benchmark datasets
that an active learning algorithm based on label-dispersion obtains excellent
results.
Related papers
- Active Learning Guided by Efficient Surrogate Learners [25.52920030051264]
Re-training a deep learning model each time a single data point receives a new label is impractical.
We introduce a new active learning algorithm that harnesses the power of a Gaussian process surrogate in conjunction with the neural network principal learner.
Our proposed model adeptly updates the surrogate learner for every new data instance, enabling it to emulate and capitalize on the continuous learning dynamics of the neural network.
arXiv Detail & Related papers (2023-01-07T01:35:25Z) - Towards Reducing Labeling Cost in Deep Object Detection [61.010693873330446]
We propose a unified framework for active learning, that considers both the uncertainty and the robustness of the detector.
Our method is able to pseudo-label the very confident predictions, suppressing a potential distribution drift.
arXiv Detail & Related papers (2021-06-22T16:53:09Z) - Low-Regret Active learning [64.36270166907788]
We develop an online learning algorithm for identifying unlabeled data points that are most informative for training.
At the core of our work is an efficient algorithm for sleeping experts that is tailored to achieve low regret on predictable (easy) instances.
arXiv Detail & Related papers (2021-04-06T22:53:45Z) - Efficacy of Bayesian Neural Networks in Active Learning [11.609770399591516]
We show that Bayesian neural networks are more efficient than ensemble based techniques in capturing uncertainty.
Our findings also reveal some key drawbacks of the ensemble techniques, which was recently shown to be more effective than Monte Carlo dropouts.
arXiv Detail & Related papers (2021-04-02T06:02:11Z) - Mutual Teaching for Graph Convolutional Networks [1.14219428942199]
Graph convolutional networks produce good predictions of unlabeled samples due to its transductive label propagation.
Since samples have different predicted confidences, we take high-confidence predictions as pseudo labels to expand the label set.
We propose a new training method named as mutual teaching, i.e., we train dual models and let them teach each other during each batch.
arXiv Detail & Related papers (2020-09-02T11:10:55Z) - Deep Active Learning in Remote Sensing for data efficient Change
Detection [26.136331738529243]
We investigate active learning in the context of deep neural network models for change detection and map updating.
In active learning, one starts from a minimal set of training examples and progressively chooses informative samples annotated by a user.
We show that active learning successfully finds highly informative samples and automatically balances the training distribution.
arXiv Detail & Related papers (2020-08-25T17:58:17Z) - On the Robustness of Active Learning [0.7340017786387767]
Active Learning is concerned with how to identify the most useful samples for a Machine Learning algorithm to be trained with.
We find that it is often applied with not enough care and domain knowledge.
We propose the new "Sum of Squared Logits" method based on the Simpson diversity index and investigate the effect of using the confusion matrix for balancing in sample selection.
arXiv Detail & Related papers (2020-06-18T09:07:23Z) - Adversarial Self-Supervised Contrastive Learning [62.17538130778111]
Existing adversarial learning approaches mostly use class labels to generate adversarial samples that lead to incorrect predictions.
We propose a novel adversarial attack for unlabeled data, which makes the model confuse the instance-level identities of the perturbed data samples.
We present a self-supervised contrastive learning framework to adversarially train a robust neural network without labeled data.
arXiv Detail & Related papers (2020-06-13T08:24:33Z) - COLAM: Co-Learning of Deep Neural Networks and Soft Labels via
Alternating Minimization [60.07531696857743]
Co-Learns DNNs and soft labels through Alternating Minimization of two objectives.
We propose COLAM framework that Co-Learns DNNs and soft labels through Alternating Minimization of two objectives.
arXiv Detail & Related papers (2020-04-26T17:50:20Z) - DivideMix: Learning with Noisy Labels as Semi-supervised Learning [111.03364864022261]
We propose DivideMix, a framework for learning with noisy labels.
Experiments on multiple benchmark datasets demonstrate substantial improvements over state-of-the-art methods.
arXiv Detail & Related papers (2020-02-18T06:20:06Z) - Learning with Out-of-Distribution Data for Audio Classification [60.48251022280506]
We show that detecting and relabelling certain OOD instances, rather than discarding them, can have a positive effect on learning.
The proposed method is shown to improve the performance of convolutional neural networks by a significant margin.
arXiv Detail & Related papers (2020-02-11T21:08:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.