Prototype Completion for Few-Shot Learning
- URL: http://arxiv.org/abs/2108.05010v1
- Date: Wed, 11 Aug 2021 03:44:00 GMT
- Title: Prototype Completion for Few-Shot Learning
- Authors: Baoquan Zhang, Xutao Li, Yunming Ye, and Shanshan Feng
- Abstract summary: Few-shot learning aims to recognize novel classes with few examples.
Pre-training based methods effectively tackle the problem by pre-training a feature extractor and then fine-tuning it through the nearest centroid based meta-learning.
We propose a novel prototype completion based meta-learning framework.
- Score: 13.63424509914303
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot learning aims to recognize novel classes with few examples.
Pre-training based methods effectively tackle the problem by pre-training a
feature extractor and then fine-tuning it through the nearest centroid based
meta-learning. However, results show that the fine-tuning step makes marginal
improvements. In this paper, 1) we figure out the reason, i.e., in the
pre-trained feature space, the base classes already form compact clusters while
novel classes spread as groups with large variances, which implies that
fine-tuning feature extractor is less meaningful; 2) instead of fine-tuning
feature extractor, we focus on estimating more representative prototypes.
Consequently, we propose a novel prototype completion based meta-learning
framework. This framework first introduces primitive knowledge (i.e.,
class-level part or attribute annotations) and extracts representative features
for seen attributes as priors. Second, a part/attribute transfer network is
designed to learn to infer the representative features for unseen attributes as
supplementary priors. Finally, a prototype completion network is devised to
learn to complete prototypes with these priors. Moreover, to avoid the
prototype completion error, we further develop a Gaussian based prototype
fusion strategy that fuses the mean-based and completed prototypes by
exploiting the unlabeled samples. Extensive experiments show that our method:
(i) obtains more accurate prototypes; (ii) achieves superior performance on
both inductive and transductive FSL settings.
Related papers
- Mixed Prototype Consistency Learning for Semi-supervised Medical Image Segmentation [0.0]
We propose the Mixed Prototype Consistency Learning (MPCL) framework, which includes a Mean Teacher and an auxiliary network.
The Mean Teacher generates prototypes for labeled and unlabeled data, while the auxiliary network produces additional prototypes for mixed data processed by CutMix.
High-quality global prototypes for each class are formed by fusing two enhanced prototypes, optimizing the distribution of hidden embeddings used in consistency learning.
arXiv Detail & Related papers (2024-04-16T16:51:12Z) - Fine-Grained Prototypes Distillation for Few-Shot Object Detection [8.795211323408513]
Few-shot object detection (FSOD) aims at extending a generic detector for novel object detection with only a few training examples.
In general, methods based on meta-learning employ an additional support branch to encode novel examples into class prototypes.
New methods are required to capture the distinctive local context for more robust novel object detection.
arXiv Detail & Related papers (2024-01-15T12:12:48Z) - Beyond Prototypes: Semantic Anchor Regularization for Better
Representation Learning [82.29761875805369]
One of the ultimate goals of representation learning is to achieve compactness within a class and well-separability between classes.
We propose a novel perspective to use pre-defined class anchors serving as feature centroid to unidirectionally guide feature learning.
The proposed Semantic Anchor Regularization (SAR) can be used in a plug-and-play manner in the existing models.
arXiv Detail & Related papers (2023-12-19T05:52:38Z) - Mixture of Gaussian-distributed Prototypes with Generative Modelling for Interpretable and Trustworthy Image Recognition [15.685927265270085]
We present a new generative paradigm to learn prototype distributions, termed as Mixture of Gaussian-distributed Prototypes (MGProto)
MGProto achieves state-of-the-art image recognition and OoD detection performances, while providing encouraging interpretability results.
arXiv Detail & Related papers (2023-11-30T11:01:37Z) - ProtoDiff: Learning to Learn Prototypical Networks by Task-Guided
Diffusion [44.805452233966534]
Prototype-based meta-learning has emerged as a powerful technique for addressing few-shot learning challenges.
We introduce ProtoDiff, a framework that gradually generates task-specific prototypes from random noise.
We conduct thorough ablation studies to demonstrate its ability to accurately capture the underlying prototype distribution.
arXiv Detail & Related papers (2023-06-26T15:26:24Z) - Evolving Semantic Prototype Improves Generative Zero-Shot Learning [73.07035277030573]
In zero-shot learning (ZSL), generative methods synthesize class-related sample features based on predefined semantic prototypes.
We observe that each class's predefined semantic prototype does not accurately match its real semantic prototype.
We propose a dynamic semantic prototype evolving (DSP) method to align the empirically predefined semantic prototypes and the real prototypes for class-related feature synthesis.
arXiv Detail & Related papers (2023-06-12T08:11:06Z) - Automatically Discovering Novel Visual Categories with Self-supervised
Prototype Learning [68.63910949916209]
This paper tackles the problem of novel category discovery (NCD), which aims to discriminate unknown categories in large-scale image collections.
We propose a novel adaptive prototype learning method consisting of two main stages: prototypical representation learning and prototypical self-training.
We conduct extensive experiments on four benchmark datasets and demonstrate the effectiveness and robustness of the proposed method with state-of-the-art performance.
arXiv Detail & Related papers (2022-08-01T16:34:33Z) - Dual Prototypical Contrastive Learning for Few-shot Semantic
Segmentation [55.339405417090084]
We propose a dual prototypical contrastive learning approach tailored to the few-shot semantic segmentation (FSS) task.
The main idea is to encourage the prototypes more discriminative by increasing inter-class distance while reducing intra-class distance in prototype feature space.
We demonstrate that the proposed dual contrastive learning approach outperforms state-of-the-art FSS methods on PASCAL-5i and COCO-20i datasets.
arXiv Detail & Related papers (2021-11-09T08:14:50Z) - Contrastive Prototype Learning with Augmented Embeddings for Few-Shot
Learning [58.2091760793799]
We propose a novel contrastive prototype learning with augmented embeddings (CPLAE) model.
With a class prototype as an anchor, CPL aims to pull the query samples of the same class closer and those of different classes further away.
Extensive experiments on several benchmarks demonstrate that our proposed CPLAE achieves new state-of-the-art.
arXiv Detail & Related papers (2021-01-23T13:22:44Z) - Prototype Completion with Primitive Knowledge for Few-Shot Learning [20.449056536438658]
Few-shot learning is a challenging task, which aims to learn a classifier for novel classes with few examples.
Pre-training based meta-learning methods effectively tackle the problem by pre-training a feature extractor and then fine-tuning it through the nearest centroid based meta-learning.
We propose a novel prototype completion based meta-learning framework.
arXiv Detail & Related papers (2020-09-10T16:09:34Z) - Prototypical Contrastive Learning of Unsupervised Representations [171.3046900127166]
Prototypical Contrastive Learning (PCL) is an unsupervised representation learning method.
PCL implicitly encodes semantic structures of the data into the learned embedding space.
PCL outperforms state-of-the-art instance-wise contrastive learning methods on multiple benchmarks.
arXiv Detail & Related papers (2020-05-11T09:53:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.