Rethinking Curriculum Learning with Incremental Labels and Adaptive
Compensation
- URL: http://arxiv.org/abs/2001.04529v3
- Date: Thu, 13 Aug 2020 16:00:02 GMT
- Title: Rethinking Curriculum Learning with Incremental Labels and Adaptive
Compensation
- Authors: Madan Ravi Ganesh and Jason J. Corso
- Abstract summary: Like humans, deep networks have been shown to learn better when samples are organized and introduced in a meaningful order or curriculum.
We propose Learning with Incremental Labels and Adaptive Compensation (LILAC), a two-phase method that incrementally increases the number of unique output labels.
- Score: 35.593312267921256
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Like humans, deep networks have been shown to learn better when samples are
organized and introduced in a meaningful order or curriculum. Conventional
curriculum learning schemes introduce samples in their order of difficulty.
This forces models to begin learning from a subset of the available data while
adding the external overhead of evaluating the difficulty of samples. In this
work, we propose Learning with Incremental Labels and Adaptive Compensation
(LILAC), a two-phase method that incrementally increases the number of unique
output labels rather than the difficulty of samples while consistently using
the entire dataset throughout training. In the first phase, Incremental Label
Introduction, we partition data into mutually exclusive subsets, one that
contains a subset of the ground-truth labels and another that contains the
remaining data attached to a pseudo-label. Throughout the training process, we
recursively reveal unseen ground-truth labels in fixed increments until all the
labels are known to the model. In the second phase, Adaptive Compensation, we
optimize the loss function using altered target vectors of previously
misclassified samples. The target vectors of such samples are modified to a
smoother distribution to help models learn better. On evaluating across three
standard image benchmarks, CIFAR-10, CIFAR-100, and STL-10, we show that LILAC
outperforms all comparable baselines. Further, we detail the importance of
pacing the introduction of new labels to a model as well as the impact of using
a smooth target vector.
Related papers
- Virtual Category Learning: A Semi-Supervised Learning Method for Dense
Prediction with Extremely Limited Labels [63.16824565919966]
This paper proposes to use confusing samples proactively without label correction.
A Virtual Category (VC) is assigned to each confusing sample in such a way that it can safely contribute to the model optimisation.
Our intriguing findings highlight the usage of VC learning in dense vision tasks.
arXiv Detail & Related papers (2023-12-02T16:23:52Z) - Rethinking Precision of Pseudo Label: Test-Time Adaptation via
Complementary Learning [10.396596055773012]
We propose a novel complementary learning approach to enhance test-time adaptation.
In test-time adaptation tasks, information from the source domain is typically unavailable.
We highlight that the risk function of complementary labels agrees with their Vanilla loss formula.
arXiv Detail & Related papers (2023-01-15T03:36:33Z) - Complementary Labels Learning with Augmented Classes [22.460256396941528]
Complementary Labels Learning (CLL) arises in many real-world tasks such as private questions classification and online learning.
We propose a novel problem setting called Complementary Labels Learning with Augmented Classes (CLLAC)
By using unlabeled data, we propose an unbiased estimator of classification risk for CLLAC, which is guaranteed to be provably consistent.
arXiv Detail & Related papers (2022-11-19T13:55:27Z) - Dense FixMatch: a simple semi-supervised learning method for pixel-wise
prediction tasks [68.36996813591425]
We propose Dense FixMatch, a simple method for online semi-supervised learning of dense and structured prediction tasks.
We enable the application of FixMatch in semi-supervised learning problems beyond image classification by adding a matching operation on the pseudo-labels.
Dense FixMatch significantly improves results compared to supervised learning using only labeled data, approaching its performance with 1/4 of the labeled samples.
arXiv Detail & Related papers (2022-10-18T15:02:51Z) - A Lagrangian Duality Approach to Active Learning [119.36233726867992]
We consider the batch active learning problem, where only a subset of the training data is labeled.
We formulate the learning problem using constrained optimization, where each constraint bounds the performance of the model on labeled samples.
We show, via numerical experiments, that our proposed approach performs similarly to or better than state-of-the-art active learning methods.
arXiv Detail & Related papers (2022-02-08T19:18:49Z) - An analysis of over-sampling labeled data in semi-supervised learning
with FixMatch [66.34968300128631]
Most semi-supervised learning methods over-sample labeled data when constructing training mini-batches.
This paper studies whether this common practice improves learning and how.
We compare it to an alternative setting where each mini-batch is uniformly sampled from all the training data, labeled or not.
arXiv Detail & Related papers (2022-01-03T12:22:26Z) - Dash: Semi-Supervised Learning with Dynamic Thresholding [72.74339790209531]
We propose a semi-supervised learning (SSL) approach that uses unlabeled examples to train models.
Our proposed approach, Dash, enjoys its adaptivity in terms of unlabeled data selection.
arXiv Detail & Related papers (2021-09-01T23:52:29Z) - Uncertainty-aware Self-training for Text Classification with Few Labels [54.13279574908808]
We study self-training as one of the earliest semi-supervised learning approaches to reduce the annotation bottleneck.
We propose an approach to improve self-training by incorporating uncertainty estimates of the underlying neural network.
We show our methods leveraging only 20-30 labeled samples per class for each task for training and for validation can perform within 3% of fully supervised pre-trained language models.
arXiv Detail & Related papers (2020-06-27T08:13:58Z) - Curriculum Labeling: Revisiting Pseudo-Labeling for Semi-Supervised
Learning [27.258077365554474]
We revisit the idea of pseudo-labeling in the context of semi-supervised learning.
Pseudo-labeling works by applying pseudo-labels to samples in the unlabeled set.
We obtain 94.91% accuracy on CIFAR-10 using only 4,000 labeled samples, and 68.87% top-1 accuracy on Imagenet-ILSVRC using only 10% of the labeled samples.
arXiv Detail & Related papers (2020-01-16T03:24:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.