Episode Adaptive Embedding Networks for Few-shot Learning
- URL: http://arxiv.org/abs/2106.09398v1
- Date: Thu, 17 Jun 2021 11:29:33 GMT
- Title: Episode Adaptive Embedding Networks for Few-shot Learning
- Authors: Fangbing Liu and Qing Wang
- Abstract summary: We propose emphEpisode Adaptive Embedding Network (EAEN) to learn episode-specific embeddings of instances.
EAEN significantly improves classification accuracy about $10%$ to $20%$ in different settings over the state-of-the-art methods.
- Score: 4.328767711595871
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Few-shot learning aims to learn a classifier using a few labelled instances
for each class. Metric-learning approaches for few-shot learning embed
instances into a high-dimensional space and conduct classification based on
distances among instance embeddings. However, such instance embeddings are
usually shared across all episodes and thus lack the discriminative power to
generalize classifiers according to episode-specific features. In this paper,
we propose a novel approach, namely \emph{Episode Adaptive Embedding Network}
(EAEN), to learn episode-specific embeddings of instances. By leveraging the
probability distributions of all instances in an episode at each channel-pixel
embedding dimension, EAEN can not only alleviate the overfitting issue
encountered in few-shot learning tasks, but also capture discriminative
features specific to an episode. To empirically verify the effectiveness and
robustness of EAEN, we have conducted extensive experiments on three widely
used benchmark datasets, under various combinations of different generic
embedding backbones and different classifiers. The results show that EAEN
significantly improves classification accuracy about $10\%$ to $20\%$ in
different settings over the state-of-the-art methods.
Related papers
- Deep Metric Learning Assisted by Intra-variance in A Semi-supervised
View of Learning [0.0]
Deep metric learning aims to construct an embedding space where samples of the same class are close to each other, while samples of different classes are far away from each other.
This paper designs a self-supervised generative assisted ranking framework that provides a semi-supervised view of intra-class variance learning scheme for typical supervised deep metric learning.
arXiv Detail & Related papers (2023-04-21T13:30:32Z) - Learning Context-aware Classifier for Semantic Segmentation [88.88198210948426]
In this paper, contextual hints are exploited via learning a context-aware classifier.
Our method is model-agnostic and can be easily applied to generic segmentation models.
With only negligible additional parameters and +2% inference time, decent performance gain has been achieved on both small and large models.
arXiv Detail & Related papers (2023-03-21T07:00:35Z) - Task-Specific Embeddings for Ante-Hoc Explainable Text Classification [6.671252951387647]
We propose an alternative training objective in which we learn task-specific embeddings of text.
Our proposed objective learns embeddings such that all texts that share the same target class label should be close together.
We present extensive experiments which show that the benefits of ante-hoc explainability and incremental learning come at no cost in overall classification accuracy.
arXiv Detail & Related papers (2022-11-30T19:56:25Z) - Adaptive Prototypical Networks [2.964978357715084]
Prototypical network for Few shot learning tries to learn an embedding function in the encoder that embeds images with similar features close to one another.
We propose an approach that intuitively pushes the embeddings of each of the classes away from the others in the meta-testing phase.
This is achieved by training the encoder network for classification using the support set samples and labels of the new task.
arXiv Detail & Related papers (2022-11-22T18:45:58Z) - Regularizing Neural Network Training via Identity-wise Discriminative
Feature Suppression [20.89979858757123]
When the number of training samples is small, or the class labels are noisy, networks tend to memorize patterns specific to individual instances to minimize the training error.
This paper explores a remedy by suppressing the network's tendency to rely on instance-specific patterns for empirical error minimisation.
arXiv Detail & Related papers (2022-09-29T05:14:56Z) - An Additive Instance-Wise Approach to Multi-class Model Interpretation [53.87578024052922]
Interpretable machine learning offers insights into what factors drive a certain prediction of a black-box system.
Existing methods mainly focus on selecting explanatory input features, which follow either locally additive or instance-wise approaches.
This work exploits the strengths of both methods and proposes a global framework for learning local explanations simultaneously for multiple target classes.
arXiv Detail & Related papers (2022-07-07T06:50:27Z) - Hyperspherical embedding for novel class classification [1.5952956981784217]
We present a constraint-based approach applied to representations in the latent space under the normalized softmax loss.
We experimentally validate the proposed approach for the classification of unseen classes on different datasets using both metric learning and the normalized softmax loss.
Our results show that not only our proposed strategy can be efficiently trained on larger set of classes, as it does not require pairwise learning, but also present better classification results than the metric learning strategies.
arXiv Detail & Related papers (2021-02-05T15:42:13Z) - Dynamic Semantic Matching and Aggregation Network for Few-shot Intent
Detection [69.2370349274216]
Few-shot Intent Detection is challenging due to the scarcity of available annotated utterances.
Semantic components are distilled from utterances via multi-head self-attention.
Our method provides a comprehensive matching measure to enhance representations of both labeled and unlabeled instances.
arXiv Detail & Related papers (2020-10-06T05:16:38Z) - K-Shot Contrastive Learning of Visual Features with Multiple Instance
Augmentations [67.46036826589467]
$K$-Shot Contrastive Learning is proposed to investigate sample variations within individual instances.
It aims to combine the advantages of inter-instance discrimination by learning discriminative features to distinguish between different instances.
Experiment results demonstrate the proposed $K$-shot contrastive learning achieves superior performances to the state-of-the-art unsupervised methods.
arXiv Detail & Related papers (2020-07-27T04:56:41Z) - Rethinking preventing class-collapsing in metric learning with
margin-based losses [81.22825616879936]
Metric learning seeks embeddings where visually similar instances are close and dissimilar instances are apart.
margin-based losses tend to project all samples of a class onto a single point in the embedding space.
We propose a simple modification to the embedding losses such that each sample selects its nearest same-class counterpart in a batch.
arXiv Detail & Related papers (2020-06-09T09:59:25Z) - Federated Learning with Only Positive Labels [71.63836379169315]
We propose a generic framework for training with only positive labels, namely Federated Averaging with Spreadout (FedAwS)
We show, both theoretically and empirically, that FedAwS can almost match the performance of conventional learning where users have access to negative labels.
arXiv Detail & Related papers (2020-04-21T23:35:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.