PTN: A Poisson Transfer Network for Semi-supervised Few-shot Learning
- URL: http://arxiv.org/abs/2012.10844v3
- Date: Thu, 1 Apr 2021 05:20:11 GMT
- Title: PTN: A Poisson Transfer Network for Semi-supervised Few-shot Learning
- Authors: Huaxi Huang, Junjie Zhang, Jian Zhang, Qiang Wu, Chang Xu
- Abstract summary: We propose a Poisson Transfer Network (PTN) to mine the unlabeled information for semi-supervised few-shot learning.
Our scheme implicitly learns the novel-class embeddings to alleviate the over-fitting problem on the few labeled data.
- Score: 21.170726615606185
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The predicament in semi-supervised few-shot learning (SSFSL) is to maximize
the value of the extra unlabeled data to boost the few-shot learner. In this
paper, we propose a Poisson Transfer Network (PTN) to mine the unlabeled
information for SSFSL from two aspects. First, the Poisson Merriman Bence Osher
(MBO) model builds a bridge for the communications between labeled and
unlabeled examples. This model serves as a more stable and informative
classifier than traditional graph-based SSFSL methods in the message-passing
process of the labels. Second, the extra unlabeled samples are employed to
transfer the knowledge from base classes to novel classes through contrastive
learning. Specifically, we force the augmented positive pairs close while push
the negative ones distant. Our contrastive transfer scheme implicitly learns
the novel-class embeddings to alleviate the over-fitting problem on the few
labeled data. Thus, we can mitigate the degeneration of embedding generality in
novel classes. Extensive experiments indicate that PTN outperforms the
state-of-the-art few-shot and SSFSL models on miniImageNet and tieredImageNet
benchmark datasets.
Related papers
- Learning to Infer from Unlabeled Data: A Semi-supervised Learning
Approach for Robust Natural Language Inference [47.293189105900524]
Natural Language Inference (NLI) aims at predicting the relation between a pair of sentences (premise and hypothesis) as entailment, contradiction or semantic independence.
Deep learning models have shown promising performance for NLI in recent years, they rely on large scale expensive human-annotated datasets.
Semi-supervised learning (SSL) is a popular technique for reducing the reliance on human annotation by leveraging unlabeled data for training.
arXiv Detail & Related papers (2022-11-05T20:34:08Z) - Self-Attention Message Passing for Contrastive Few-Shot Learning [2.1485350418225244]
Unsupervised few-shot learning is the pursuit of bridging this gap between machines and humans.
We propose a novel self-attention based message passing contrastive learning approach (coined as SAMP-CLR) for U-FSL pre-training.
We also propose an optimal transport (OT) based fine-tuning strategy (we call OpT-Tune) to efficiently induce task awareness into our novel end-to-end unsupervised few-shot classification framework (SAMPTransfer)
arXiv Detail & Related papers (2022-10-12T15:57:44Z) - OpenLDN: Learning to Discover Novel Classes for Open-World
Semi-Supervised Learning [110.40285771431687]
Semi-supervised learning (SSL) is one of the dominant approaches to address the annotation bottleneck of supervised learning.
Recent SSL methods can effectively leverage a large repository of unlabeled data to improve performance while relying on a small set of labeled data.
This work introduces OpenLDN that utilizes a pairwise similarity loss to discover novel classes.
arXiv Detail & Related papers (2022-07-05T18:51:05Z) - On Non-Random Missing Labels in Semi-Supervised Learning [114.62655062520425]
Semi-Supervised Learning (SSL) is fundamentally a missing label problem.
We explicitly incorporate "class" into SSL.
Our method not only significantly outperforms existing baselines but also surpasses other label bias removal SSL methods.
arXiv Detail & Related papers (2022-06-29T22:01:29Z) - GCT: Graph Co-Training for Semi-Supervised Few-Shot Learning [46.07923908746946]
We propose a novel label prediction method, Isolated Graph Learning (IGL)
IGL introduces the Laplacian operator to encode the raw data to graph space, which helps reduce the dependence on features when classifying.
Second, we propose Graph Co-Training (GCT) to tackle this challenge from a multi-modal fusion perspective.
arXiv Detail & Related papers (2022-03-15T09:13:35Z) - L2B: Learning to Bootstrap Robust Models for Combating Label Noise [52.02335367411447]
This paper introduces a simple and effective method, named Learning to Bootstrap (L2B)
It enables models to bootstrap themselves using their own predictions without being adversely affected by erroneous pseudo-labels.
It achieves this by dynamically adjusting the importance weight between real observed and generated labels, as well as between different samples through meta-learning.
arXiv Detail & Related papers (2022-02-09T05:57:08Z) - A Strong Baseline for Semi-Supervised Incremental Few-Shot Learning [54.617688468341704]
Few-shot learning aims to learn models that generalize to novel classes with limited training samples.
We propose a novel paradigm containing two parts: (1) a well-designed meta-training algorithm for mitigating ambiguity between base and novel classes caused by unreliable pseudo labels and (2) a model adaptation mechanism to learn discriminative features for novel classes while preserving base knowledge using few labeled and all the unlabeled data.
arXiv Detail & Related papers (2021-10-21T13:25:52Z) - Few-shot Learning via Dependency Maximization and Instance Discriminant
Analysis [21.8311401851523]
We study the few-shot learning problem, where a model learns to recognize new objects with extremely few labeled data per category.
We propose a simple approach to exploit unlabeled data accompanying the few-shot task for improving few-shot performance.
arXiv Detail & Related papers (2021-09-07T02:19:01Z) - Information Bottleneck Constrained Latent Bidirectional Embedding for
Zero-Shot Learning [59.58381904522967]
We propose a novel embedding based generative model with a tight visual-semantic coupling constraint.
We learn a unified latent space that calibrates the embedded parametric distributions of both visual and semantic spaces.
Our method can be easily extended to transductive ZSL setting by generating labels for unseen images.
arXiv Detail & Related papers (2020-09-16T03:54:12Z) - Instance Credibility Inference for Few-Shot Learning [45.577880041135785]
Few-shot learning aims to recognize new objects with extremely limited training data for each category.
This paper presents a simple statistical approach, dubbed Instance Credibility Inference (ICI) to exploit the distribution support of unlabeled instances for few-shot learning.
Our simple approach can establish new state-of-the-arts on four widely used few-shot learning benchmark datasets.
arXiv Detail & Related papers (2020-03-26T12:01:15Z) - TAFSSL: Task-Adaptive Feature Sub-Space Learning for few-shot
classification [50.358839666165764]
We show that the Task-Adaptive Feature Sub-Space Learning (TAFSSL) can significantly boost the performance in Few-Shot Learning scenarios.
Specifically, we show that on the challenging miniImageNet and tieredImageNet benchmarks, TAFSSL can improve the current state-of-the-art in both transductive and semi-supervised FSL settings by more than $5%$.
arXiv Detail & Related papers (2020-03-14T16:59:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.