Learning the Prediction Distribution for Semi-Supervised Learning with
Normalising Flows
- URL: http://arxiv.org/abs/2007.02745v1
- Date: Mon, 6 Jul 2020 13:36:00 GMT
- Title: Learning the Prediction Distribution for Semi-Supervised Learning with
Normalising Flows
- Authors: Ivana Bala\v{z}evi\'c, Carl Allen, Timothy Hospedales
- Abstract summary: Impressive results have been achieved in semi-supervised learning (SSL) for image classification, nearing fully supervised performance.
We propose a probabilistically principled general approach to SSL that considers the distribution over label predictions.
We demonstrate the general applicability of this approach on a range of computer vision tasks with varying output complexity.
- Score: 6.789370732159177
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As data volumes continue to grow, the labelling process increasingly becomes
a bottleneck, creating demand for methods that leverage information from
unlabelled data. Impressive results have been achieved in semi-supervised
learning (SSL) for image classification, nearing fully supervised performance,
with only a fraction of the data labelled. In this work, we propose a
probabilistically principled general approach to SSL that considers the
distribution over label predictions, for labels of different complexity, from
"one-hot" vectors to binary vectors and images. Our method regularises an
underlying supervised model, using a normalising flow that learns the posterior
distribution over predictions for labelled data, to serve as a prior over the
predictions on unlabelled data. We demonstrate the general applicability of
this approach on a range of computer vision tasks with varying output
complexity: classification, attribute prediction and image-to-image
translation.
Related papers
- Continuous Contrastive Learning for Long-Tailed Semi-Supervised Recognition [50.61991746981703]
Current state-of-the-art LTSSL approaches rely on high-quality pseudo-labels for large-scale unlabeled data.
This paper introduces a novel probabilistic framework that unifies various recent proposals in long-tail learning.
We introduce a continuous contrastive learning method, CCL, extending our framework to unlabeled data using reliable and smoothed pseudo-labels.
arXiv Detail & Related papers (2024-10-08T15:06:10Z) - Exploring Beyond Logits: Hierarchical Dynamic Labeling Based on Embeddings for Semi-Supervised Classification [49.09505771145326]
We propose a Hierarchical Dynamic Labeling (HDL) algorithm that does not depend on model predictions and utilizes image embeddings to generate sample labels.
Our approach has the potential to change the paradigm of pseudo-label generation in semi-supervised learning.
arXiv Detail & Related papers (2024-04-26T06:00:27Z) - Semi-Supervised Learning for hyperspectral images by non parametrically
predicting view assignment [25.198550162904713]
Hyperspectral image (HSI) classification is gaining a lot of momentum in present time because of high inherent spectral information within the images.
Recently, to effectively train the deep learning models with minimal labelled samples, the unlabeled samples are also being leveraged in self-supervised and semi-supervised setting.
In this work, we leverage the idea of semi-supervised learning to assist the discriminative self-supervised pretraining of the models.
arXiv Detail & Related papers (2023-06-19T14:13:56Z) - Dense FixMatch: a simple semi-supervised learning method for pixel-wise
prediction tasks [68.36996813591425]
We propose Dense FixMatch, a simple method for online semi-supervised learning of dense and structured prediction tasks.
We enable the application of FixMatch in semi-supervised learning problems beyond image classification by adding a matching operation on the pseudo-labels.
Dense FixMatch significantly improves results compared to supervised learning using only labeled data, approaching its performance with 1/4 of the labeled samples.
arXiv Detail & Related papers (2022-10-18T15:02:51Z) - Out-distribution aware Self-training in an Open World Setting [62.19882458285749]
We leverage unlabeled data in an open world setting to further improve prediction performance.
We introduce out-distribution aware self-training, which includes a careful sample selection strategy.
Our classifiers are by design out-distribution aware and can thus distinguish task-related inputs from unrelated ones.
arXiv Detail & Related papers (2020-12-21T12:25:04Z) - Semi-Supervised Models via Data Augmentationfor Classifying Interactive
Affective Responses [85.04362095899656]
We present semi-supervised models with data augmentation (SMDA), a semi-supervised text classification system to classify interactive affective responses.
For labeled sentences, we performed data augmentations to uniform the label distributions and computed supervised loss during training process.
For unlabeled sentences, we explored self-training by regarding low-entropy predictions over unlabeled sentences as pseudo labels.
arXiv Detail & Related papers (2020-04-23T05:02:31Z) - Semi-Supervised Learning with Normalizing Flows [54.376602201489995]
FlowGMM is an end-to-end approach to generative semi supervised learning with normalizing flows.
We show promising results on a wide range of applications, including AG-News and Yahoo Answers text data.
arXiv Detail & Related papers (2019-12-30T17:36:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.