BaseTransformers: Attention over base data-points for One Shot Learning
- URL: http://arxiv.org/abs/2210.02476v1
- Date: Wed, 5 Oct 2022 18:00:24 GMT
- Title: BaseTransformers: Attention over base data-points for One Shot Learning
- Authors: Mayug Maniparambil, Kevin McGuinness, Noel O'Connor
- Abstract summary: Few shot classification aims to learn to recognize novel categories using only limited samples per category.
Most current few shot methods use a base dataset rich in labeled examples to train an encoder that is used for obtaining representations of support instances for novel classes.
In this paper we propose to make use of the well-trained feature representations of the base dataset that are closest to each support instance to improve its representation during meta-test time.
- Score: 6.708284461619745
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Few shot classification aims to learn to recognize novel categories using
only limited samples per category. Most current few shot methods use a base
dataset rich in labeled examples to train an encoder that is used for obtaining
representations of support instances for novel classes. Since the test
instances are from a distribution different to the base distribution, their
feature representations are of poor quality, degrading performance. In this
paper we propose to make use of the well-trained feature representations of the
base dataset that are closest to each support instance to improve its
representation during meta-test time. To this end, we propose BaseTransformers,
that attends to the most relevant regions of the base dataset feature space and
improves support instance representations. Experiments on three benchmark data
sets show that our method works well for several backbones and achieves
state-of-the-art results in the inductive one shot setting. Code is available
at github.com/mayug/BaseTransformers
Related papers
- Rethinking Few-shot 3D Point Cloud Semantic Segmentation [62.80639841429669]
This paper revisits few-shot 3D point cloud semantic segmentation (FS-PCS)
We focus on two significant issues in the state-of-the-art: foreground leakage and sparse point distribution.
To address these issues, we introduce a standardized FS-PCS setting, upon which a new benchmark is built.
arXiv Detail & Related papers (2024-03-01T15:14:47Z) - Few and Fewer: Learning Better from Few Examples Using Fewer Base
Classes [12.742052888217543]
Fine-tuning is ineffective for few-shot learning, since the target dataset contains only a handful of examples.
This paper investigates whether better features for the target dataset can be obtained by training on fewer base classes.
To our knowledge, this is the first demonstration that fine-tuning on a subset of carefully selected base classes can significantly improve few-shot learning.
arXiv Detail & Related papers (2024-01-29T01:52:49Z) - Few-Shot Object Detection via Variational Feature Aggregation [32.34871873486389]
We propose a meta-learning framework with two novel feature aggregation schemes.
We first present a Class-Agnostic Aggregation (CAA) method, where the query and support features can be aggregated regardless of their categories.
We then propose a Variational Feature Aggregation (VFA) method, which encodes support examples into class-level support features.
arXiv Detail & Related papers (2023-01-31T04:58:21Z) - Prediction Calibration for Generalized Few-shot Semantic Segmentation [101.69940565204816]
Generalized Few-shot Semantic (GFSS) aims to segment each image pixel into either base classes with abundant training examples or novel classes with only a handful of (e.g., 1-5) training images per class.
We build a cross-attention module that guides the classifier's final prediction using the fused multi-level features.
Our PCN outperforms the state-the-art alternatives by large margins.
arXiv Detail & Related papers (2022-10-15T13:30:12Z) - A Closer Look at Few-Shot Video Classification: A New Baseline and
Benchmark [33.86872697028233]
We present an in-depth study on few-shot video classification by making three contributions.
First, we perform a consistent comparative study on the existing metric-based methods to figure out their limitations in representation learning.
Second, we discover that there is a high correlation between the novel action class and the ImageNet object class, which is problematic in the few-shot recognition setting.
Third, we present a new benchmark with more base data to facilitate future few-shot video classification without pre-training.
arXiv Detail & Related papers (2021-10-24T06:01:46Z) - Revisiting Deep Local Descriptor for Improved Few-Shot Classification [56.74552164206737]
We show how one can improve the quality of embeddings by leveraging textbfDense textbfClassification and textbfAttentive textbfPooling.
We suggest to pool feature maps by applying attentive pooling instead of the widely used global average pooling (GAP) to prepare embeddings for few-shot classification.
arXiv Detail & Related papers (2021-03-30T00:48:28Z) - A Universal Representation Transformer Layer for Few-Shot Image
Classification [43.31379752656756]
Few-shot classification aims to recognize unseen classes when presented with only a small number of samples.
We consider the problem of multi-domain few-shot image classification, where unseen classes and examples come from diverse data sources.
Here, we propose a Universal Representation Transformer layer, that meta-learns to leverage universal features for few-shot classification.
arXiv Detail & Related papers (2020-06-21T03:08:00Z) - UniT: Unified Knowledge Transfer for Any-shot Object Detection and
Segmentation [52.487469544343305]
Methods for object detection and segmentation rely on large scale instance-level annotations for training.
We propose an intuitive and unified semi-supervised model that is applicable to a range of supervision.
arXiv Detail & Related papers (2020-06-12T22:45:47Z) - Selecting Relevant Features from a Multi-domain Representation for
Few-shot Classification [91.67977602992657]
We propose a new strategy based on feature selection, which is both simpler and more effective than previous feature adaptation approaches.
We show that a simple non-parametric classifier built on top of such features produces high accuracy and generalizes to domains never seen during training.
arXiv Detail & Related papers (2020-03-20T15:44:17Z) - Frustratingly Simple Few-Shot Object Detection [98.42824677627581]
We find that fine-tuning only the last layer of existing detectors on rare classes is crucial to the few-shot object detection task.
Such a simple approach outperforms the meta-learning methods by roughly 220 points on current benchmarks.
arXiv Detail & Related papers (2020-03-16T00:29:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.