Self-Attention Message Passing for Contrastive Few-Shot Learning
- URL: http://arxiv.org/abs/2210.06339v1
- Date: Wed, 12 Oct 2022 15:57:44 GMT
- Title: Self-Attention Message Passing for Contrastive Few-Shot Learning
- Authors: Ojas Kishorkumar Shirekar, Anuj Singh, Hadi Jamali-Rad
- Abstract summary: Unsupervised few-shot learning is the pursuit of bridging this gap between machines and humans.
We propose a novel self-attention based message passing contrastive learning approach (coined as SAMP-CLR) for U-FSL pre-training.
We also propose an optimal transport (OT) based fine-tuning strategy (we call OpT-Tune) to efficiently induce task awareness into our novel end-to-end unsupervised few-shot classification framework (SAMPTransfer)
- Score: 2.1485350418225244
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Humans have a unique ability to learn new representations from just a handful
of examples with little to no supervision. Deep learning models, however,
require an abundance of data and supervision to perform at a satisfactory
level. Unsupervised few-shot learning (U-FSL) is the pursuit of bridging this
gap between machines and humans. Inspired by the capacity of graph neural
networks (GNNs) in discovering complex inter-sample relationships, we propose a
novel self-attention based message passing contrastive learning approach
(coined as SAMP-CLR) for U-FSL pre-training. We also propose an optimal
transport (OT) based fine-tuning strategy (we call OpT-Tune) to efficiently
induce task awareness into our novel end-to-end unsupervised few-shot
classification framework (SAMPTransfer). Our extensive experimental results
corroborate the efficacy of SAMPTransfer in a variety of downstream few-shot
classification scenarios, setting a new state-of-the-art for U-FSL on both
miniImagenet and tieredImagenet benchmarks, offering up to 7%+ and 5%+
improvements, respectively. Our further investigations also confirm that
SAMPTransfer remains on-par with some supervised baselines on miniImagenet and
outperforms all existing U-FSL baselines in a challenging cross-domain
scenario. Our code can be found in our GitHub repository at
https://github.com/ojss/SAMPTransfer/.
Related papers
- Intra-task Mutual Attention based Vision Transformer for Few-Shot Learning [12.5354658533836]
Humans possess remarkable ability to accurately classify new, unseen images after being exposed to only a few examples.
For artificial neural network models, determining the most relevant features for distinguishing between two images with limited samples presents a challenge.
We propose an intra-task mutual attention method for few-shot learning, that involves splitting the support and query samples into patches.
arXiv Detail & Related papers (2024-05-06T02:02:57Z) - BECLR: Batch Enhanced Contrastive Few-Shot Learning [1.450405446885067]
Unsupervised few-shot learning aspires to bridge this gap by discarding the reliance on annotations at training time.
We propose a novel Dynamic Clustered mEmory (DyCE) module to promote a highly separable latent representation space.
We then tackle the, somehow overlooked yet critical, issue of sample bias at the few-shot inference stage.
arXiv Detail & Related papers (2024-02-04T10:52:43Z) - MOCA: Self-supervised Representation Learning by Predicting Masked Online Codebook Assignments [72.6405488990753]
Self-supervised learning can be used for mitigating the greedy needs of Vision Transformer networks.
We propose a single-stage and standalone method, MOCA, which unifies both desired properties.
We achieve new state-of-the-art results on low-shot settings and strong experimental results in various evaluation protocols.
arXiv Detail & Related papers (2023-07-18T15:46:20Z) - Siamese Prototypical Contrastive Learning [24.794022951873156]
Contrastive Self-supervised Learning (CSL) is a practical solution that learns meaningful visual representations from massive data in an unsupervised approach.
In this paper, we tackle this problem by introducing a simple but effective contrastive learning framework.
The key insight is to employ siamese-style metric loss to match intra-prototype features, while increasing the distance between inter-prototype features.
arXiv Detail & Related papers (2022-08-18T13:25:30Z) - Learning to Win Lottery Tickets in BERT Transfer via Task-agnostic Mask
Training [55.43088293183165]
Recent studies show that pre-trained language models (PLMs) like BERT contain matchingworks that have similar transfer learning performance as the original PLM.
In this paper, we find that the BERTworks have even more potential than these studies have shown.
We train binary masks over model weights on the pre-training tasks, with the aim of preserving the universal transferability of the subnetwork.
arXiv Detail & Related papers (2022-04-24T08:42:47Z) - Pushing the Limits of Simple Pipelines for Few-Shot Learning: External
Data and Fine-Tuning Make a Difference [74.80730361332711]
Few-shot learning is an important and topical problem in computer vision.
We show that a simple transformer-based pipeline yields surprisingly good performance on standard benchmarks.
arXiv Detail & Related papers (2022-04-15T02:55:58Z) - Trash to Treasure: Harvesting OOD Data with Cross-Modal Matching for
Open-Set Semi-Supervised Learning [101.28281124670647]
Open-set semi-supervised learning (open-set SSL) investigates a challenging but practical scenario where out-of-distribution (OOD) samples are contained in the unlabeled data.
We propose a novel training mechanism that could effectively exploit the presence of OOD data for enhanced feature learning.
Our approach substantially lifts the performance on open-set SSL and outperforms the state-of-the-art by a large margin.
arXiv Detail & Related papers (2021-08-12T09:14:44Z) - Learning Invariant Representations across Domains and Tasks [81.30046935430791]
We propose a novel Task Adaptation Network (TAN) to solve this unsupervised task transfer problem.
In addition to learning transferable features via domain-adversarial training, we propose a novel task semantic adaptor that uses the learning-to-learn strategy to adapt the task semantics.
TAN significantly increases the recall and F1 score by 5.0% and 7.8% compared to recently strong baselines.
arXiv Detail & Related papers (2021-03-03T11:18:43Z) - PTN: A Poisson Transfer Network for Semi-supervised Few-shot Learning [21.170726615606185]
We propose a Poisson Transfer Network (PTN) to mine the unlabeled information for semi-supervised few-shot learning.
Our scheme implicitly learns the novel-class embeddings to alleviate the over-fitting problem on the few labeled data.
arXiv Detail & Related papers (2020-12-20T04:44:37Z) - TAFSSL: Task-Adaptive Feature Sub-Space Learning for few-shot
classification [50.358839666165764]
We show that the Task-Adaptive Feature Sub-Space Learning (TAFSSL) can significantly boost the performance in Few-Shot Learning scenarios.
Specifically, we show that on the challenging miniImageNet and tieredImageNet benchmarks, TAFSSL can improve the current state-of-the-art in both transductive and semi-supervised FSL settings by more than $5%$.
arXiv Detail & Related papers (2020-03-14T16:59:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.