Soft Curriculum for Learning Conditional GANs with Noisy-Labeled and
Uncurated Unlabeled Data
- URL: http://arxiv.org/abs/2307.08319v1
- Date: Mon, 17 Jul 2023 08:31:59 GMT
- Title: Soft Curriculum for Learning Conditional GANs with Noisy-Labeled and
Uncurated Unlabeled Data
- Authors: Kai Katsumata, Duc Minh Vo, Tatsuya Harada, Hideki Nakayama
- Abstract summary: We introduce a novel conditional image generation framework that accepts noisy-labeled and uncurated data during training.
We propose soft curriculum learning, which assigns instance-wise weights for adversarial training while assigning new labels for unlabeled data.
Our experiments show that our approach outperforms existing semi-supervised and label-noise robust methods in terms of both quantitative and qualitative performance.
- Score: 70.25049762295193
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Label-noise or curated unlabeled data is used to compensate for the
assumption of clean labeled data in training the conditional generative
adversarial network; however, satisfying such an extended assumption is
occasionally laborious or impractical. As a step towards generative modeling
accessible to everyone, we introduce a novel conditional image generation
framework that accepts noisy-labeled and uncurated unlabeled data during
training: (i) closed-set and open-set label noise in labeled data and (ii)
closed-set and open-set unlabeled data. To combat it, we propose soft
curriculum learning, which assigns instance-wise weights for adversarial
training while assigning new labels for unlabeled data and correcting wrong
labels for labeled data. Unlike popular curriculum learning, which uses a
threshold to pick the training samples, our soft curriculum controls the effect
of each training instance by using the weights predicted by the auxiliary
classifier, resulting in the preservation of useful samples while ignoring
harmful ones. Our experiments show that our approach outperforms existing
semi-supervised and label-noise robust methods in terms of both quantitative
and qualitative performance. In particular, the proposed approach is able to
match the performance of (semi-) supervised GANs even with less than half the
labeled data.
Related papers
- Semi-Supervised Variational Adversarial Active Learning via Learning to Rank and Agreement-Based Pseudo Labeling [6.771578432805963]
Active learning aims to alleviate the amount of labor involved in data labeling by automating the selection of unlabeled samples.
We introduce novel techniques that significantly improve the use of abundant unlabeled data during training.
We demonstrate the superior performance of our approach over the state of the art on various image classification and segmentation benchmark datasets.
arXiv Detail & Related papers (2024-08-23T00:35:07Z) - Debiased Pseudo Labeling in Self-Training [77.83549261035277]
Deep neural networks achieve remarkable performances on a wide range of tasks with the aid of large-scale labeled datasets.
To mitigate the requirement for labeled data, self-training is widely used in both academia and industry by pseudo labeling on readily-available unlabeled data.
We propose Debiased, in which the generation and utilization of pseudo labels are decoupled by two independent heads.
arXiv Detail & Related papers (2022-02-15T02:14:33Z) - Dash: Semi-Supervised Learning with Dynamic Thresholding [72.74339790209531]
We propose a semi-supervised learning (SSL) approach that uses unlabeled examples to train models.
Our proposed approach, Dash, enjoys its adaptivity in terms of unlabeled data selection.
arXiv Detail & Related papers (2021-09-01T23:52:29Z) - Active label cleaning: Improving dataset quality under resource
constraints [13.716577886649018]
Imperfections in data annotation, known as label noise, are detrimental to the training of machine learning models.
This work advocates for a data-driven approach to prioritising samples for re-annotation.
We rank instances according to estimated label correctness and labelling difficulty of each sample, and introduce a simulation framework to evaluate relabelling efficacy.
arXiv Detail & Related papers (2021-09-01T19:03:57Z) - OpenCoS: Contrastive Semi-supervised Learning for Handling Open-set
Unlabeled Data [65.19205979542305]
Unlabeled data may include out-of-class samples in practice.
OpenCoS is a method for handling this realistic semi-supervised learning scenario.
arXiv Detail & Related papers (2021-06-29T06:10:05Z) - Instance Correction for Learning with Open-set Noisy Labels [145.06552420999986]
We use the sample selection approach to handle open-set noisy labels.
The discarded data are seen to be mislabeled and do not participate in training.
We modify the instances of discarded data to make predictions for the discarded data consistent with given labels.
arXiv Detail & Related papers (2021-06-01T13:05:55Z) - A Novel Perspective for Positive-Unlabeled Learning via Noisy Labels [49.990938653249415]
This research presents a methodology that assigns initial pseudo-labels to unlabeled data which is used as noisy-labeled data, and trains a deep neural network using the noisy-labeled data.
Experimental results demonstrate that the proposed method significantly outperforms the state-of-the-art methods on several benchmark datasets.
arXiv Detail & Related papers (2021-03-08T11:46:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.