OpenCoS: Contrastive Semi-supervised Learning for Handling Open-set
Unlabeled Data
- URL: http://arxiv.org/abs/2107.08943v1
- Date: Tue, 29 Jun 2021 06:10:05 GMT
- Title: OpenCoS: Contrastive Semi-supervised Learning for Handling Open-set
Unlabeled Data
- Authors: Jongjin Park, Sukmin Yun, Jongheon Jeong, Jinwoo Shin
- Abstract summary: Unlabeled data may include out-of-class samples in practice.
OpenCoS is a method for handling this realistic semi-supervised learning scenario.
- Score: 65.19205979542305
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern semi-supervised learning methods conventionally assume both labeled
and unlabeled data have the same class distribution. However, unlabeled data
may include out-of-class samples in practice; those that cannot have one-hot
encoded labels from a closed-set of classes in label data, i.e., unlabeled data
is an open-set. In this paper, we introduce OpenCoS, a method for handling this
realistic semi-supervised learning scenario based on a recent framework of
contrastive learning. One of our key findings is that out-of-class samples in
the unlabeled dataset can be identified effectively via (unsupervised)
contrastive learning. OpenCoS utilizes this information to overcome the failure
modes in the existing state-of-the-art semi-supervised methods, e.g.,
ReMixMatch or FixMatch. It further improves the semi-supervised performance by
utilizing soft- and pseudo-labels on open-set unlabeled data, learned from
contrastive learning. Our extensive experimental results show the effectiveness
of OpenCoS, fixing the state-of-the-art semi-supervised methods to be suitable
for diverse scenarios involving open-set unlabeled data.
Related papers
- FlatMatch: Bridging Labeled Data and Unlabeled Data with Cross-Sharpness
for Semi-Supervised Learning [73.13448439554497]
Semi-Supervised Learning (SSL) has been an effective way to leverage abundant unlabeled data with extremely scarce labeled data.
Most SSL methods are commonly based on instance-wise consistency between different data transformations.
We propose FlatMatch which minimizes a cross-sharpness measure to ensure consistent learning performance between the two datasets.
arXiv Detail & Related papers (2023-10-25T06:57:59Z) - Soft Curriculum for Learning Conditional GANs with Noisy-Labeled and
Uncurated Unlabeled Data [70.25049762295193]
We introduce a novel conditional image generation framework that accepts noisy-labeled and uncurated data during training.
We propose soft curriculum learning, which assigns instance-wise weights for adversarial training while assigning new labels for unlabeled data.
Our experiments show that our approach outperforms existing semi-supervised and label-noise robust methods in terms of both quantitative and qualitative performance.
arXiv Detail & Related papers (2023-07-17T08:31:59Z) - Are labels informative in semi-supervised learning? -- Estimating and
leveraging the missing-data mechanism [4.675583319625962]
Semi-supervised learning is a powerful technique for leveraging unlabeled data to improve machine learning models.
It can be affected by the presence of informative'' labels, which occur when some classes are more likely to be labeled than others.
We propose a novel approach to address this issue by estimating the missing-data mechanism and using inverse propensity weighting to debias any SSL algorithm.
arXiv Detail & Related papers (2023-02-15T09:18:46Z) - OSSGAN: Open-Set Semi-Supervised Image Generation [26.67298827670573]
We introduce a challenging training scheme of conditional GANs, called open-set semi-supervised image generation.
OSSGAN provides decision clues to the discriminator on the basis of whether an unlabeled image belongs to one or none of the classes of interest.
The results of experiments on Tiny ImageNet and ImageNet show notable improvements over supervised BigGAN and semi-supervised methods.
arXiv Detail & Related papers (2022-04-29T17:26:09Z) - GuidedMix-Net: Semi-supervised Semantic Segmentation by Using Labeled
Images as Reference [90.5402652758316]
We propose a novel method for semi-supervised semantic segmentation named GuidedMix-Net.
It uses labeled information to guide the learning of unlabeled instances.
It achieves competitive segmentation accuracy and significantly improves the mIoU by +7$%$ compared to previous approaches.
arXiv Detail & Related papers (2021-12-28T06:48:03Z) - GuidedMix-Net: Learning to Improve Pseudo Masks Using Labeled Images as
Reference [153.354332374204]
We propose a novel method for semi-supervised semantic segmentation named GuidedMix-Net.
We first introduce a feature alignment objective between labeled and unlabeled data to capture potentially similar image pairs.
MITrans is shown to be a powerful knowledge module for further progressive refining features of unlabeled data.
Along with supervised learning for labeled data, the prediction of unlabeled data is jointly learned with the generated pseudo masks.
arXiv Detail & Related papers (2021-06-29T02:48:45Z) - OpenMix: Reviving Known Knowledge for Discovering Novel Visual
Categories in An Open World [127.64076228829606]
We introduce OpenMix to mix the unlabeled examples from an open set and the labeled examples from known classes.
OpenMix helps to prevent the model from overfitting on unlabeled samples that may be assigned with wrong pseudo-labels.
arXiv Detail & Related papers (2020-04-12T05:52:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.