Generating Representative Samples for Few-Shot Classification
- URL: http://arxiv.org/abs/2205.02918v1
- Date: Thu, 5 May 2022 20:58:33 GMT
- Title: Generating Representative Samples for Few-Shot Classification
- Authors: Jingyi Xu and Hieu Le
- Abstract summary: Few-shot learning aims to learn new categories with a few visual samples per class.
Few-shot class representations are often biased due to data scarcity.
We generate visual samples based on semantic embeddings using a conditional variational autoencoder model.
- Score: 8.62483598990205
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Few-shot learning (FSL) aims to learn new categories with a few visual
samples per class. Few-shot class representations are often biased due to data
scarcity. To mitigate this issue, we propose to generate visual samples based
on semantic embeddings using a conditional variational autoencoder (CVAE)
model. We train this CVAE model on base classes and use it to generate features
for novel classes. More importantly, we guide this VAE to strictly generate
representative samples by removing non-representative samples from the base
training set when training the CVAE model. We show that this training scheme
enhances the representativeness of the generated samples and therefore,
improves the few-shot classification results. Experimental results show that
our method improves three FSL baseline methods by substantial margins,
achieving state-of-the-art few-shot classification performance on miniImageNet
and tieredImageNet datasets for both 1-shot and 5-shot settings. Code is
available at: https://github.com/cvlab-stonybrook/fsl-rsvae.
Related papers
- FreeSeg-Diff: Training-Free Open-Vocabulary Segmentation with Diffusion Models [56.71672127740099]
We focus on the task of image segmentation, which is traditionally solved by training models on closed-vocabulary datasets.
We leverage different and relatively small-sized, open-source foundation models for zero-shot open-vocabulary segmentation.
Our approach (dubbed FreeSeg-Diff), which does not rely on any training, outperforms many training-based approaches on both Pascal VOC and COCO datasets.
arXiv Detail & Related papers (2024-03-29T10:38:25Z) - FSL-Rectifier: Rectify Outliers in Few-Shot Learning via Test-Time Augmentation [7.477118370563593]
Few-shot-learning (FSL) commonly requires a model to identify images (queries) that belong to classes unseen during training.
We generate additional test-class samples by combining original samples with suitable train-class samples via a generative image combiner.
We obtain averaged features via an augmentor, which leads to more typical representations through the averaging.
arXiv Detail & Related papers (2024-02-28T12:37:30Z) - Diversified in-domain synthesis with efficient fine-tuning for few-shot
classification [64.86872227580866]
Few-shot image classification aims to learn an image classifier using only a small set of labeled examples per class.
We propose DISEF, a novel approach which addresses the generalization challenge in few-shot learning using synthetic data.
We validate our method in ten different benchmarks, consistently outperforming baselines and establishing a new state-of-the-art for few-shot classification.
arXiv Detail & Related papers (2023-12-05T17:18:09Z) - Ref-Diff: Zero-shot Referring Image Segmentation with Generative Models [68.73086826874733]
We introduce a novel Referring Diffusional segmentor (Ref-Diff) for referring image segmentation.
We demonstrate that without a proposal generator, a generative model alone can achieve comparable performance to existing SOTA weakly-supervised models.
This indicates that generative models are also beneficial for this task and can complement discriminative models for better referring segmentation.
arXiv Detail & Related papers (2023-08-31T14:55:30Z) - Image-free Classifier Injection for Zero-Shot Classification [72.66409483088995]
Zero-shot learning models achieve remarkable results on image classification for samples from classes that were not seen during training.
We aim to equip pre-trained models with zero-shot classification capabilities without the use of image data.
We achieve this with our proposed Image-free Injection with Semantics (ICIS)
arXiv Detail & Related papers (2023-08-21T09:56:48Z) - Kernel Relative-prototype Spectral Filtering for Few-shot Learning [3.2091741098687696]
Few-shot learning performs classification tasks and regression tasks on scarce samples.
In this paper, we propose a framework of spectral filtering (shrinkage) for measuring the difference between query samples and prototypes.
arXiv Detail & Related papers (2022-07-24T07:53:27Z) - Forgetting Data from Pre-trained GANs [28.326418377665345]
We investigate how to post-edit a model after training so that it forgets certain kinds of samples.
We provide three different algorithms for GANs that differ on how the samples to be forgotten are described.
Our algorithms are capable of forgetting data while retaining high generation quality at a fraction of the cost of full re-training.
arXiv Detail & Related papers (2022-06-29T03:46:16Z) - Ortho-Shot: Low Displacement Rank Regularization with Data Augmentation
for Few-Shot Learning [23.465747123791772]
In few-shot classification, the primary goal is to learn representations that generalize well for novel classes.
We propose an efficient low displacement rank (LDR) regularization strategy termed Ortho-Shot.
arXiv Detail & Related papers (2021-10-18T14:58:36Z) - Contrastive Prototype Learning with Augmented Embeddings for Few-Shot
Learning [58.2091760793799]
We propose a novel contrastive prototype learning with augmented embeddings (CPLAE) model.
With a class prototype as an anchor, CPL aims to pull the query samples of the same class closer and those of different classes further away.
Extensive experiments on several benchmarks demonstrate that our proposed CPLAE achieves new state-of-the-art.
arXiv Detail & Related papers (2021-01-23T13:22:44Z) - Few-shot Classification via Adaptive Attention [93.06105498633492]
We propose a novel few-shot learning method via optimizing and fast adapting the query sample representation based on very few reference samples.
As demonstrated experimentally, the proposed model achieves state-of-the-art classification results on various benchmark few-shot classification and fine-grained recognition datasets.
arXiv Detail & Related papers (2020-08-06T05:52:59Z) - Interpretable Time-series Classification on Few-shot Samples [27.05851877375113]
This paper proposes an interpretable neural-based framework, namely textitDual Prototypical Shapelet Networks (DPSN) for few-shot time-series classification.
DPSN interprets the model from dual granularity: 1) global overview using representative time series samples, and 2) local highlights using discriminative shapelets.
We have derived 18 few-shot TSC datasets from public benchmark datasets and evaluated the proposed method by comparing with baselines.
arXiv Detail & Related papers (2020-06-03T03:47:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.