Unsupervised Semantic Aggregation and Deformable Template Matching for
Semi-Supervised Learning
- URL: http://arxiv.org/abs/2010.05517v1
- Date: Mon, 12 Oct 2020 08:17:56 GMT
- Title: Unsupervised Semantic Aggregation and Deformable Template Matching for
Semi-Supervised Learning
- Authors: Tao Han, Junyu Gao, Yuan Yuan, Qi Wang
- Abstract summary: Unsupervised semantic aggregation based on T-MI loss is explored to generate semantic labels for unlabeled data.
A feature pool that stores the labeled samples is dynamically updated to assign proxy labels for unlabeled data.
Experiments and analysis validate that USADTM achieves top performance.
- Score: 34.560447389853614
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unlabeled data learning has attracted considerable attention recently.
However, it is still elusive to extract the expected high-level semantic
feature with mere unsupervised learning. In the meantime, semi-supervised
learning (SSL) demonstrates a promising future in leveraging few samples. In
this paper, we combine both to propose an Unsupervised Semantic Aggregation and
Deformable Template Matching (USADTM) framework for SSL, which strives to
improve the classification performance with few labeled data and then reduce
the cost in data annotating. Specifically, unsupervised semantic aggregation
based on Triplet Mutual Information (T-MI) loss is explored to generate
semantic labels for unlabeled data. Then the semantic labels are aligned to the
actual class by the supervision of labeled data. Furthermore, a feature pool
that stores the labeled samples is dynamically updated to assign proxy labels
for unlabeled data, which are used as targets for cross-entropy minimization.
Extensive experiments and analysis across four standard semi-supervised
learning benchmarks validate that USADTM achieves top performance (e.g.,
90.46$\%$ accuracy on CIFAR-10 with 40 labels and 95.20$\%$ accuracy with 250
labels). The code is released at https://github.com/taohan10200/USADTM.
Related papers
- FlatMatch: Bridging Labeled Data and Unlabeled Data with Cross-Sharpness
for Semi-Supervised Learning [73.13448439554497]
Semi-Supervised Learning (SSL) has been an effective way to leverage abundant unlabeled data with extremely scarce labeled data.
Most SSL methods are commonly based on instance-wise consistency between different data transformations.
We propose FlatMatch which minimizes a cross-sharpness measure to ensure consistent learning performance between the two datasets.
arXiv Detail & Related papers (2023-10-25T06:57:59Z) - Dist-PU: Positive-Unlabeled Learning from a Label Distribution
Perspective [89.5370481649529]
We propose a label distribution perspective for PU learning in this paper.
Motivated by this, we propose to pursue the label distribution consistency between predicted and ground-truth label distributions.
Experiments on three benchmark datasets validate the effectiveness of the proposed method.
arXiv Detail & Related papers (2022-12-06T07:38:29Z) - Contrastive Credibility Propagation for Reliable Semi-Supervised Learning [6.014538614447467]
We propose Contrastive Credibility Propagation (CCP) for deep SSL via iterative transductive pseudo-label refinement.
CCP unifies semi-supervised learning and noisy label learning for the goal of reliably outperforming a supervised baseline in any data scenario.
arXiv Detail & Related papers (2022-11-17T23:01:47Z) - Pseudo-Labeling Based Practical Semi-Supervised Meta-Training for Few-Shot Learning [93.63638405586354]
We propose a simple and effective meta-training framework, called pseudo-labeling based meta-learning (PLML)
Firstly, we train a classifier via common semi-supervised learning (SSL) and use it to obtain the pseudo-labels of unlabeled data.
We build few-shot tasks from labeled and pseudo-labeled data and design a novel finetuning method with feature smoothing and noise suppression.
arXiv Detail & Related papers (2022-07-14T10:53:53Z) - AuxMix: Semi-Supervised Learning with Unconstrained Unlabeled Data [6.633920993895286]
We show that state-of-the-art SSL algorithms suffer a degradation in performance in the presence of unlabeled auxiliary data.
We propose AuxMix, an algorithm that leverages self-supervised learning tasks to learn generic features in order to mask auxiliary data that are not semantically similar to the labeled set.
arXiv Detail & Related papers (2022-06-14T16:25:20Z) - Label Noise-Resistant Mean Teaching for Weakly Supervised Fake News
Detection [93.6222609806278]
We propose a novel label noise-resistant mean teaching approach (LNMT) for weakly supervised fake news detection.
LNMT leverages unlabeled news and feedback comments of users to enlarge the amount of training data.
LNMT establishes a mean teacher framework equipped with label propagation and label reliability estimation.
arXiv Detail & Related papers (2022-06-10T16:01:58Z) - CLS: Cross Labeling Supervision for Semi-Supervised Learning [9.929229055862491]
Cross Labeling Supervision ( CLS) is a framework that generalizes the typical pseudo-labeling process.
CLS allows the creation of both pseudo and complementary labels to support both positive and negative learning.
arXiv Detail & Related papers (2022-02-17T08:09:40Z) - Unsupervised Selective Labeling for More Effective Semi-Supervised
Learning [46.414510522978425]
unsupervised selective labeling consistently improves SSL methods over state-of-the-art active learning given labeled data.
Our work sets a new standard for practical and efficient SSL.
arXiv Detail & Related papers (2021-10-06T18:25:50Z) - Dash: Semi-Supervised Learning with Dynamic Thresholding [72.74339790209531]
We propose a semi-supervised learning (SSL) approach that uses unlabeled examples to train models.
Our proposed approach, Dash, enjoys its adaptivity in terms of unlabeled data selection.
arXiv Detail & Related papers (2021-09-01T23:52:29Z) - Self-Tuning for Data-Efficient Deep Learning [75.34320911480008]
Self-Tuning is a novel approach to enable data-efficient deep learning.
It unifies the exploration of labeled and unlabeled data and the transfer of a pre-trained model.
It outperforms its SSL and TL counterparts on five tasks by sharp margins.
arXiv Detail & Related papers (2021-02-25T14:56:19Z) - Analysis of label noise in graph-based semi-supervised learning [2.4366811507669124]
In machine learning, one must acquire labels to help supervise a model that will be able to generalize to unseen data.
It is often the case that most of our data is unlabeled.
Semi-supervised learning (SSL) alleviates that by making strong assumptions about the relation between the labels and the input data distribution.
arXiv Detail & Related papers (2020-09-27T22:13:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.