Meta-Generating Deep Attentive Metric for Few-shot Classification
- URL: http://arxiv.org/abs/2012.01641v1
- Date: Thu, 3 Dec 2020 02:07:43 GMT
- Title: Meta-Generating Deep Attentive Metric for Few-shot Classification
- Authors: Lei Zhang, Fei Zhou, Wei Wei and Yanning Zhang
- Abstract summary: We present a novel deep metric meta-generation method to generate a specific metric for a new few-shot learning task.
In this study, we structure the metric using a three-layer deep attentive network that is flexible enough to produce a discriminative metric for each task.
We gain surprisingly obvious performance improvement over state-of-the-art competitors, especially in the challenging cases.
- Score: 53.07108067253006
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning to generate a task-aware base learner proves a promising direction
to deal with few-shot learning (FSL) problem. Existing methods mainly focus on
generating an embedding model utilized with a fixed metric (eg, cosine
distance) for nearest neighbour classification or directly generating a linear
classier. However, due to the limited discriminative capacity of such a simple
metric or classifier, these methods fail to generalize to challenging cases
appropriately. To mitigate this problem, we present a novel deep metric
meta-generation method that turns to an orthogonal direction, ie, learning to
adaptively generate a specific metric for a new FSL task based on the task
description (eg, a few labelled samples). In this study, we structure the
metric using a three-layer deep attentive network that is flexible enough to
produce a discriminative metric for each task. Moreover, different from
existing methods that utilize an uni-modal weight distribution conditioned on
labelled samples for network generation, the proposed meta-learner establishes
a multi-modal weight distribution conditioned on cross-class sample pairs using
a tailored variational autoencoder, which can separately capture the specific
inter-class discrepancy statistics for each class and jointly embed the
statistics for all classes into metric generation. By doing this, the generated
metric can be appropriately adapted to a new FSL task with pleasing
generalization performance. To demonstrate this, we test the proposed method on
four benchmark FSL datasets and gain surprisingly obvious performance
improvement over state-of-the-art competitors, especially in the challenging
cases, eg, improve the accuracy from 26.14% to 46.69% in the 20-way 1-shot task
on miniImageNet, while improve the accuracy from 45.2% to 68.72% in the 5-way
1-shot task on FC100. Code is available: https://github.com/NWPUZhoufei/DAM.
Related papers
- Hyperspherical Classification with Dynamic Label-to-Prototype Assignment [5.978350039412277]
We present a simple yet effective method to optimize the category assigned to each prototype during the training.
We solve this optimization using a sequential combination of gradient descent and Bipartide matching.
Our method outperforms its competitors by 1.22% accuracy on CIFAR-100, and 2.15% on ImageNet-200 using a metric space dimension half of the size of its competitors.
arXiv Detail & Related papers (2024-03-25T17:01:34Z) - Rethinking Classifier Re-Training in Long-Tailed Recognition: A Simple
Logits Retargeting Approach [102.0769560460338]
We develop a simple logits approach (LORT) without the requirement of prior knowledge of the number of samples per class.
Our method achieves state-of-the-art performance on various imbalanced datasets, including CIFAR100-LT, ImageNet-LT, and iNaturalist 2018.
arXiv Detail & Related papers (2024-03-01T03:27:08Z) - An Adaptive Plug-and-Play Network for Few-Shot Learning [12.023266104119289]
Few-shot learning requires a model to classify new samples after learning from only a few samples.
Deep networks and complex metrics tend to induce overfitting, making it difficult to further improve the performance.
We propose plug-and-play model-adaptive resizer (MAR) and adaptive similarity metric (ASM) without any other losses.
arXiv Detail & Related papers (2023-02-18T13:25:04Z) - Intra-class Adaptive Augmentation with Neighbor Correction for Deep
Metric Learning [99.14132861655223]
We propose a novel intra-class adaptive augmentation (IAA) framework for deep metric learning.
We reasonably estimate intra-class variations for every class and generate adaptive synthetic samples to support hard samples mining.
Our method significantly improves and outperforms the state-of-the-art methods on retrieval performances by 3%-6%.
arXiv Detail & Related papers (2022-11-29T14:52:38Z) - Self-Adaptive Label Augmentation for Semi-supervised Few-shot
Classification [121.63992191386502]
Few-shot classification aims to learn a model that can generalize well to new tasks when only a few labeled samples are available.
We propose a semi-supervised few-shot classification method that assigns an appropriate label to each unlabeled sample by a manually defined metric.
A major novelty of SALA is the task-adaptive metric, which can learn the metric adaptively for different tasks in an end-to-end fashion.
arXiv Detail & Related papers (2022-06-16T13:14:03Z) - Lightweight Conditional Model Extrapolation for Streaming Data under
Class-Prior Shift [27.806085423595334]
We introduce LIMES, a new method for learning with non-stationary streaming data.
We learn a single set of model parameters from which a specific classifier for any specific data distribution is derived.
Experiments on a set of exemplary tasks using Twitter data show that LIMES achieves higher accuracy than alternative approaches.
arXiv Detail & Related papers (2022-06-10T15:19:52Z) - Exploring Complementary Strengths of Invariant and Equivariant
Representations for Few-Shot Learning [96.75889543560497]
In many real-world problems, collecting a large number of labeled samples is infeasible.
Few-shot learning is the dominant approach to address this issue, where the objective is to quickly adapt to novel categories in presence of a limited number of samples.
We propose a novel training mechanism that simultaneously enforces equivariance and invariance to a general set of geometric transformations.
arXiv Detail & Related papers (2021-03-01T21:14:33Z) - Meta-Regularization by Enforcing Mutual-Exclusiveness [0.8057006406834467]
We propose a regularization technique for meta-learning models that gives the model designer more control over the information flow during meta-training.
Our proposed regularization function shows an accuracy boost of $sim$ $36%$ on the Omniglot dataset.
arXiv Detail & Related papers (2021-01-24T22:57:19Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.