Boosting Few-Shot Learning With Adaptive Margin Loss
- URL: http://arxiv.org/abs/2005.13826v1
- Date: Thu, 28 May 2020 07:58:41 GMT
- Title: Boosting Few-Shot Learning With Adaptive Margin Loss
- Authors: Aoxue Li and Weiran Huang and Xu Lan and Jiashi Feng and Zhenguo Li
and Liwei Wang
- Abstract summary: This paper proposes an adaptive margin principle to improve the generalization ability of metric-based meta-learning approaches for few-shot learning problems.
Extensive experiments demonstrate that the proposed method can boost the performance of current metric-based meta-learning approaches.
- Score: 109.03665126222619
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot learning (FSL) has attracted increasing attention in recent years
but remains challenging, due to the intrinsic difficulty in learning to
generalize from a few examples. This paper proposes an adaptive margin
principle to improve the generalization ability of metric-based meta-learning
approaches for few-shot learning problems. Specifically, we first develop a
class-relevant additive margin loss, where semantic similarity between each
pair of classes is considered to separate samples in the feature embedding
space from similar classes. Further, we incorporate the semantic context among
all classes in a sampled training task and develop a task-relevant additive
margin loss to better distinguish samples from different classes. Our adaptive
margin method can be easily extended to a more realistic generalized FSL
setting. Extensive experiments demonstrate that the proposed method can boost
the performance of current metric-based meta-learning approaches, under both
the standard FSL and generalized FSL settings.
Related papers
- BECLR: Batch Enhanced Contrastive Few-Shot Learning [1.450405446885067]
Unsupervised few-shot learning aspires to bridge this gap by discarding the reliance on annotations at training time.
We propose a novel Dynamic Clustered mEmory (DyCE) module to promote a highly separable latent representation space.
We then tackle the, somehow overlooked yet critical, issue of sample bias at the few-shot inference stage.
arXiv Detail & Related papers (2024-02-04T10:52:43Z) - Intra-class Adaptive Augmentation with Neighbor Correction for Deep
Metric Learning [99.14132861655223]
We propose a novel intra-class adaptive augmentation (IAA) framework for deep metric learning.
We reasonably estimate intra-class variations for every class and generate adaptive synthetic samples to support hard samples mining.
Our method significantly improves and outperforms the state-of-the-art methods on retrieval performances by 3%-6%.
arXiv Detail & Related papers (2022-11-29T14:52:38Z) - Semantics-driven Attentive Few-shot Learning over Clean and Noisy
Samples [0.0]
We aim to train meta-learner models that can leverage prior semantic knowledge about novel classes to guide the classifier synthesis process.
In particular, we propose semantically-conditioned feature attention and sample attention mechanisms that estimate the importance of representation dimensions and training instances.
arXiv Detail & Related papers (2022-01-09T16:16:23Z) - A Strong Baseline for Semi-Supervised Incremental Few-Shot Learning [54.617688468341704]
Few-shot learning aims to learn models that generalize to novel classes with limited training samples.
We propose a novel paradigm containing two parts: (1) a well-designed meta-training algorithm for mitigating ambiguity between base and novel classes caused by unreliable pseudo labels and (2) a model adaptation mechanism to learn discriminative features for novel classes while preserving base knowledge using few labeled and all the unlabeled data.
arXiv Detail & Related papers (2021-10-21T13:25:52Z) - The Role of Global Labels in Few-Shot Classification and How to Infer
Them [55.64429518100676]
Few-shot learning is a central problem in meta-learning, where learners must quickly adapt to new tasks.
We propose Meta Label Learning (MeLa), a novel algorithm that infers global labels and obtains robust few-shot models via standard classification.
arXiv Detail & Related papers (2021-08-09T14:07:46Z) - ReSSL: Relational Self-Supervised Learning with Weak Augmentation [68.47096022526927]
Self-supervised learning has achieved great success in learning visual representations without data annotations.
We introduce a novel relational SSL paradigm that learns representations by modeling the relationship between different instances.
Our proposed ReSSL significantly outperforms the previous state-of-the-art algorithms in terms of both performance and training efficiency.
arXiv Detail & Related papers (2021-07-20T06:53:07Z) - Exploring Complementary Strengths of Invariant and Equivariant
Representations for Few-Shot Learning [96.75889543560497]
In many real-world problems, collecting a large number of labeled samples is infeasible.
Few-shot learning is the dominant approach to address this issue, where the objective is to quickly adapt to novel categories in presence of a limited number of samples.
We propose a novel training mechanism that simultaneously enforces equivariance and invariance to a general set of geometric transformations.
arXiv Detail & Related papers (2021-03-01T21:14:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.