On the Importance of Distractors for Few-Shot Classification
- URL: http://arxiv.org/abs/2109.09883v1
- Date: Mon, 20 Sep 2021 23:35:56 GMT
- Title: On the Importance of Distractors for Few-Shot Classification
- Authors: Rajshekhar Das, Yu-Xiong Wang, Jos\'eM.F. Moura
- Abstract summary: Few-shot classification aims at classifying categories of a novel task by learning from just a few (typically, 1 to 5) labelled examples.
An effective approach to few-shot classification involves a prior model trained on a large-sample base domain, which is then finetuned over the novel few-shot task to yield generalizable representations.
We propose a new finetuning approach based on contrastive learning that reuses unlabelled examples from the base domain in the form of distractors.
- Score: 13.486661391097387
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot classification aims at classifying categories of a novel task by
learning from just a few (typically, 1 to 5) labelled examples. An effective
approach to few-shot classification involves a prior model trained on a
large-sample base domain, which is then finetuned over the novel few-shot task
to yield generalizable representations. However, task-specific finetuning is
prone to overfitting due to the lack of enough training examples. To alleviate
this issue, we propose a new finetuning approach based on contrastive learning
that reuses unlabelled examples from the base domain in the form of
distractors. Unlike the nature of unlabelled data used in prior works,
distractors belong to classes that do not overlap with the novel categories. We
demonstrate for the first time that inclusion of such distractors can
significantly boost few-shot generalization. Our technical novelty includes a
stochastic pairing of examples sharing the same category in the few-shot task
and a weighting term that controls the relative influence of task-specific
negatives and distractors. An important aspect of our finetuning objective is
that it is agnostic to distractor labels and hence applicable to various base
domain settings. Compared to state-of-the-art approaches, our method shows
accuracy gains of up to $12\%$ in cross-domain and up to $5\%$ in unsupervised
prior-learning settings.
Related papers
- Liberating Seen Classes: Boosting Few-Shot and Zero-Shot Text Classification via Anchor Generation and Classification Reframing [38.84431954053434]
Few-shot and zero-shot text classification aim to recognize samples from novel classes with limited labeled samples or no labeled samples at all.
We propose a simple and effective strategy for few-shot and zero-shot text classification.
arXiv Detail & Related papers (2024-05-06T15:38:32Z) - Capturing Perspectives of Crowdsourced Annotators in Subjective Learning Tasks [9.110872603799839]
Supervised classification heavily depends on datasets annotated by humans.
In subjective tasks such as toxicity classification, these annotations often exhibit low agreement among raters.
In this work, we propose textbfAnnotator Awares for Texts (AART) for subjective classification tasks.
arXiv Detail & Related papers (2023-11-16T10:18:32Z) - Few-shot Forgery Detection via Guided Adversarial Interpolation [56.59499187594308]
Existing forgery detection methods suffer from significant performance drops when applied to unseen novel forgery approaches.
We propose Guided Adversarial Interpolation (GAI) to overcome the few-shot forgery detection problem.
Our method is validated to be robust to choices of majority and minority forgery approaches.
arXiv Detail & Related papers (2022-04-12T16:05:10Z) - Bridging Non Co-occurrence with Unlabeled In-the-wild Data for
Incremental Object Detection [56.22467011292147]
Several incremental learning methods are proposed to mitigate catastrophic forgetting for object detection.
Despite the effectiveness, these methods require co-occurrence of the unlabeled base classes in the training data of the novel classes.
We propose the use of unlabeled in-the-wild data to bridge the non-occurrence caused by the missing base classes during the training of additional novel classes.
arXiv Detail & Related papers (2021-10-28T10:57:25Z) - Instance-Level Relative Saliency Ranking with Graph Reasoning [126.09138829920627]
We present a novel unified model to segment salient instances and infer relative saliency rank order.
A novel loss function is also proposed to effectively train the saliency ranking branch.
experimental results demonstrate that our proposed model is more effective than previous methods.
arXiv Detail & Related papers (2021-07-08T13:10:42Z) - Few-shot Action Recognition with Prototype-centered Attentive Learning [88.10852114988829]
Prototype-centered Attentive Learning (PAL) model composed of two novel components.
First, a prototype-centered contrastive learning loss is introduced to complement the conventional query-centered learning objective.
Second, PAL integrates a attentive hybrid learning mechanism that can minimize the negative impacts of outliers.
arXiv Detail & Related papers (2021-01-20T11:48:12Z) - Shot in the Dark: Few-Shot Learning with No Base-Class Labels [32.96824710484196]
We show that off-the-shelf self-supervised learning outperforms transductive few-shot methods by 3.9% for 5-shot accuracy on miniImageNet.
This motivates us to examine more carefully the role of features learned through self-supervision in few-shot learning.
arXiv Detail & Related papers (2020-10-06T02:05:27Z) - Cooperative Bi-path Metric for Few-shot Learning [50.98891758059389]
We make two contributions to investigate the few-shot classification problem.
We report a simple and effective baseline trained on base classes in the way of traditional supervised learning.
We propose a cooperative bi-path metric for classification, which leverages the correlations between base classes and novel classes to further improve the accuracy.
arXiv Detail & Related papers (2020-08-10T11:28:52Z) - Towards Cross-Granularity Few-Shot Learning: Coarse-to-Fine
Pseudo-Labeling with Visual-Semantic Meta-Embedding [13.063136901934865]
Few-shot learning aims at rapidly adapting to novel categories with only a handful of samples at test time.
In this paper, we advance the few-shot classification paradigm towards a more challenging scenario, i.e., cross-granularity few-shot classification.
We approximate the fine-grained data distribution by greedy clustering of each coarse-class into pseudo-fine-classes according to the similarity of image embeddings.
arXiv Detail & Related papers (2020-07-11T03:44:21Z) - Any-Shot Object Detection [81.88153407655334]
'Any-shot detection' is where totally unseen and few-shot categories can simultaneously co-occur during inference.
We propose a unified any-shot detection model, that can concurrently learn to detect both zero-shot and few-shot object classes.
Our framework can also be used solely for Zero-shot detection and Few-shot detection tasks.
arXiv Detail & Related papers (2020-03-16T03:43:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.