MutexMatch: Semi-supervised Learning with Mutex-based Consistency
Regularization
- URL: http://arxiv.org/abs/2203.14316v1
- Date: Sun, 27 Mar 2022 14:28:16 GMT
- Title: MutexMatch: Semi-supervised Learning with Mutex-based Consistency
Regularization
- Authors: Yue Duan, Zhen Zhao, Lei Qi, Lei Wang, Luping Zhou, Yinghuan Shi, Yang
Gao
- Abstract summary: We propose a mutex-based consistency regularization, namely Mutex, to utilize low-confidence samples in a novel way.
MutexMatch achieves superior performance on multiple benchmark datasets, i.e., CIFAR-10, CIFAR-100, SVHN, STL-10, and mini-ImageNet.
- Score: 36.019086181632005
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The core issue in semi-supervised learning (SSL) lies in how to effectively
leverage unlabeled data, whereas most existing methods tend to put a great
emphasis on the utilization of high-confidence samples yet seldom fully explore
the usage of low-confidence samples. In this paper, we aim to utilize
low-confidence samples in a novel way with our proposed mutex-based consistency
regularization, namely MutexMatch. Specifically, the high-confidence samples
are required to exactly predict "what it is" by conventional True-Positive
Classifier, while the low-confidence samples are employed to achieve a simpler
goal -- to predict with ease "what it is not" by True-Negative Classifier. In
this sense, we not only mitigate the pseudo-labeling errors but also make full
use of the low-confidence unlabeled data by consistency of dissimilarity
degree. MutexMatch achieves superior performance on multiple benchmark
datasets, i.e., CIFAR-10, CIFAR-100, SVHN, STL-10, and mini-ImageNet. More
importantly, our method further shows superiority when the amount of labeled
data is scarce, e.g., 92.23% accuracy with only 20 labeled data on CIFAR-10.
Code has been released at https://github.com/NJUyued/MutexMatch4SSL.
Related papers
- VCC-INFUSE: Towards Accurate and Efficient Selection of Unlabeled Examples in Semi-supervised Learning [1.0007051676618415]
Existing methods fail to utilize unlabeled data effectively and efficiently.
We propose two methods: Variational Confidence (VCC) and Influence-Function-based Unlabeled Sample Elimination (INFUSE)
Our methods are effective in multiple datasets and settings, reducing errors classification rates and saving training time.
arXiv Detail & Related papers (2024-04-18T06:59:40Z) - Shrinking Class Space for Enhanced Certainty in Semi-Supervised Learning [59.44422468242455]
We propose a novel method dubbed ShrinkMatch to learn uncertain samples.
For each uncertain sample, it adaptively seeks a shrunk class space, which merely contains the original top-1 class.
We then impose a consistency regularization between a pair of strongly and weakly augmented samples in the shrunk space to strive for discriminative representations.
arXiv Detail & Related papers (2023-08-13T14:05:24Z) - Boosting Semi-Supervised Learning by Exploiting All Unlabeled Data [21.6350640726058]
Semi-supervised learning (SSL) has attracted enormous attention due to its vast potential of mitigating the dependence on large labeled datasets.
We propose two novel techniques: Entropy Meaning Loss (EML) and Adaptive Negative Learning (ANL)
We integrate these techniques with FixMatch, and develop a simple yet powerful framework called FullMatch.
arXiv Detail & Related papers (2023-03-20T12:44:11Z) - FreeMatch: Self-adaptive Thresholding for Semi-supervised Learning [46.95063831057502]
We propose emphFreeMatch to define and adjust the confidence threshold in a self-adaptive manner according to the model's learning status.
FreeMatch achieves textbf5.78%, textbf13.59%, and textbf1.28% error rate reduction over the latest state-of-the-art method FlexMatch on CIFAR-10 with 1 label per class.
arXiv Detail & Related papers (2022-05-15T10:07:52Z) - Towards Semi-Supervised Deep Facial Expression Recognition with An
Adaptive Confidence Margin [92.76372026435858]
We learn an Adaptive Confidence Margin (Ada-CM) to fully leverage all unlabeled data for semi-supervised deep facial expression recognition.
All unlabeled samples are partitioned into two subsets by comparing their confidence scores with the adaptively learned confidence margin.
Our method achieves state-of-the-art performance, especially surpassing fully-supervised baselines in a semi-supervised manner.
arXiv Detail & Related papers (2022-03-23T11:43:29Z) - OpenMatch: Open-set Consistency Regularization for Semi-supervised
Learning with Outliers [71.08167292329028]
We propose a novel Open-set Semi-Supervised Learning (OSSL) approach called OpenMatch.
OpenMatch unifies FixMatch with novelty detection based on one-vs-all (OVA) classifiers.
It achieves state-of-the-art performance on three datasets, and even outperforms a fully supervised model in detecting outliers unseen in unlabeled data on CIFAR10.
arXiv Detail & Related papers (2021-05-28T23:57:15Z) - In Defense of Pseudo-Labeling: An Uncertainty-Aware Pseudo-label
Selection Framework for Semi-Supervised Learning [53.1047775185362]
Pseudo-labeling (PL) is a general SSL approach that does not have this constraint but performs relatively poorly in its original formulation.
We argue that PL underperforms due to the erroneous high confidence predictions from poorly calibrated models.
We propose an uncertainty-aware pseudo-label selection (UPS) framework which improves pseudo labeling accuracy by drastically reducing the amount of noise encountered in the training process.
arXiv Detail & Related papers (2021-01-15T23:29:57Z) - FixMatch: Simplifying Semi-Supervised Learning with Consistency and
Confidence [93.91751021370638]
Semi-supervised learning (SSL) provides an effective means of leveraging unlabeled data to improve a model's performance.
In this paper, we demonstrate the power of a simple combination of two common SSL methods: consistency regularization and pseudo-labeling.
Our algorithm, FixMatch, first generates pseudo-labels using the model's predictions on weakly-augmented unlabeled images.
arXiv Detail & Related papers (2020-01-21T18:32:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.