Learning Primitive-aware Discriminative Representations for Few-shot
Learning
- URL: http://arxiv.org/abs/2208.09717v2
- Date: Wed, 14 Jun 2023 16:54:31 GMT
- Title: Learning Primitive-aware Discriminative Representations for Few-shot
Learning
- Authors: Jianpeng Yang, Yuhang Niu, Xuemei Xie, Guangming Shi
- Abstract summary: Few-shot learning aims to learn a classifier that can be easily adapted to recognize novel classes with only a few labeled examples.
We propose a Primitive Mining and Reasoning Network (PMRN) to learn primitive-aware representations.
Our method achieves state-of-the-art results on six standard benchmarks.
- Score: 28.17404445820028
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot learning (FSL) aims to learn a classifier that can be easily adapted
to recognize novel classes with only a few labeled examples. Some recent work
about FSL has yielded promising classification performance, where the
image-level feature is used to calculate the similarity among samples for
classification. However, the image-level feature ignores abundant fine-grained
and structural in-formation of objects that may be transferable and consistent
between seen and unseen classes. How can humans easily identify novel classes
with several sam-ples? Some study from cognitive science argues that humans can
recognize novel categories through primitives. Although base and novel
categories are non-overlapping, they can share some primitives in common.
Inspired by above re-search, we propose a Primitive Mining and Reasoning
Network (PMRN) to learn primitive-aware representations based on metric-based
FSL model. Concretely, we first add Self-supervision Jigsaw task (SSJ) for
feature extractor parallelly, guiding the model to encode visual pattern
corresponding to object parts into fea-ture channels. To further mine
discriminative representations, an Adaptive Chan-nel Grouping (ACG) method is
applied to cluster and weight spatially and se-mantically related visual
patterns to generate a group of visual primitives. To fur-ther enhance the
discriminability and transferability of primitives, we propose a visual
primitive Correlation Reasoning Network (CRN) based on graph convolu-tional
network to learn abundant structural information and internal correlation among
primitives. Finally, a primitive-level metric is conducted for classification
in a meta-task based on episodic training strategy. Extensive experiments show
that our method achieves state-of-the-art results on six standard benchmarks.
Related papers
- Preview-based Category Contrastive Learning for Knowledge Distillation [53.551002781828146]
We propose a novel preview-based category contrastive learning method for knowledge distillation (PCKD)
It first distills the structural knowledge of both instance-level feature correspondence and the relation between instance features and category centers.
It can explicitly optimize the category representation and explore the distinct correlation between representations of instances and categories.
arXiv Detail & Related papers (2024-10-18T03:31:00Z) - Queryable Prototype Multiple Instance Learning with Vision-Language Models for Incremental Whole Slide Image Classification [10.667645628712542]
This paper proposes the first Vision-Language-based framework with Queryable Prototype Multiple Instance Learning (QPMIL-VL) specially designed for incremental WSI classification.
experiments on four TCGA datasets demonstrate that our QPMIL-VL framework is effective for incremental WSI classification.
arXiv Detail & Related papers (2024-10-14T14:49:34Z) - Few-Shot Class-Incremental Learning via Training-Free Prototype
Calibration [67.69532794049445]
We find a tendency for existing methods to misclassify the samples of new classes into base classes, which leads to the poor performance of new classes.
We propose a simple yet effective Training-frEE calibratioN (TEEN) strategy to enhance the discriminability of new classes.
arXiv Detail & Related papers (2023-12-08T18:24:08Z) - ECKPN: Explicit Class Knowledge Propagation Network for Transductive
Few-shot Learning [53.09923823663554]
Class-level knowledge can be easily learned by humans from just a handful of samples.
We propose an Explicit Class Knowledge Propagation Network (ECKPN) to address this problem.
We conduct extensive experiments on four few-shot classification benchmarks, and the experimental results show that the proposed ECKPN significantly outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2021-06-16T02:29:43Z) - GAN for Vision, KG for Relation: a Two-stage Deep Network for Zero-shot
Action Recognition [33.23662792742078]
We propose a two-stage deep neural network for zero-shot action recognition.
In the sampling stage, we utilize a generative adversarial networks (GAN) trained by action features and word vectors of seen classes.
In the classification stage, we construct a knowledge graph based on the relationship between word vectors of action classes and related objects.
arXiv Detail & Related papers (2021-05-25T09:34:42Z) - Semi-Supervised Few-Shot Classification with Deep Invertible Hybrid
Models [4.189643331553922]
We propose a deep invertible hybrid model which integrates discriminative and generative learning at a latent space level for semi-supervised few-shot classification.
Our main originality lies in our integration of these components at a latent space level, which is effective in preventing overfitting.
arXiv Detail & Related papers (2021-05-22T05:55:16Z) - Adaptive Prototypical Networks with Label Words and Joint Representation
Learning for Few-Shot Relation Classification [17.237331828747006]
This work focuses on few-shot relation classification (FSRC)
We propose an adaptive mixture mechanism to add label words to the representation of the class prototype.
Experiments have been conducted on FewRel under different few-shot (FS) settings.
arXiv Detail & Related papers (2021-01-10T11:25:42Z) - Learning and Evaluating Representations for Deep One-class
Classification [59.095144932794646]
We present a two-stage framework for deep one-class classification.
We first learn self-supervised representations from one-class data, and then build one-class classifiers on learned representations.
In experiments, we demonstrate state-of-the-art performance on visual domain one-class classification benchmarks.
arXiv Detail & Related papers (2020-11-04T23:33:41Z) - Region Comparison Network for Interpretable Few-shot Image
Classification [97.97902360117368]
Few-shot image classification has been proposed to effectively use only a limited number of labeled examples to train models for new classes.
We propose a metric learning based method named Region Comparison Network (RCN), which is able to reveal how few-shot learning works.
We also present a new way to generalize the interpretability from the level of tasks to categories.
arXiv Detail & Related papers (2020-09-08T07:29:05Z) - Dataset Bias in Few-shot Image Recognition [57.25445414402398]
We first investigate the impact of transferable capabilities learned from base categories.
Second, we investigate performance differences on different datasets from dataset structures and different few-shot learning methods.
arXiv Detail & Related papers (2020-08-18T14:46:23Z) - Compositional Few-Shot Recognition with Primitive Discovery and
Enhancing [43.478770119996184]
Few-shot learning aims at recognizing novel classes given only few training samples.
Humans can easily recognize novel classes with only few samples.
We propose an approach to learn a feature representation composed of important primitives.
arXiv Detail & Related papers (2020-05-12T21:01:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.