Mixture-based Feature Space Learning for Few-shot Image Classification
- URL: http://arxiv.org/abs/2011.11872v2
- Date: Tue, 17 Aug 2021 17:44:41 GMT
- Title: Mixture-based Feature Space Learning for Few-shot Image Classification
- Authors: Arman Afrasiyabi, Jean-Fran\c{c}ois Lalonde, Christian Gagn\'e
- Abstract summary: We propose to model base classes with mixture models by simultaneously training the feature extractor and learning the mixture model parameters in an online manner.
Results in a richer and more discriminative feature space which can be employed to classify novel examples from very few samples.
- Score: 6.574517227976925
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce Mixture-based Feature Space Learning (MixtFSL) for obtaining a
rich and robust feature representation in the context of few-shot image
classification. Previous works have proposed to model each base class either
with a single point or with a mixture model by relying on offline clustering
algorithms. In contrast, we propose to model base classes with mixture models
by simultaneously training the feature extractor and learning the mixture model
parameters in an online manner. This results in a richer and more
discriminative feature space which can be employed to classify novel examples
from very few samples. Two main stages are proposed to train the MixtFSL model.
First, the multimodal mixtures for each base class and the feature extractor
parameters are learned using a combination of two loss functions. Second, the
resulting network and mixture models are progressively refined through a
leader-follower learning procedure, which uses the current estimate as a
"target" network. This target network is used to make a consistent assignment
of instances to mixture components, which increases performance and stabilizes
training. The effectiveness of our end-to-end feature space learning approach
is demonstrated with extensive experiments on four standard datasets and four
backbones. Notably, we demonstrate that when we combine our robust
representation with recent alignment-based approaches, we achieve new
state-of-the-art results in the inductive setting, with an absolute accuracy
for 5-shot classification of 82.45 on miniImageNet, 88.20 with tieredImageNet,
and 60.70 in FC100 using the ResNet-12 backbone.
Related papers
- Adversarial AutoMixup [50.1874436169571]
We propose AdAutomixup, an adversarial automatic mixup augmentation approach.
It generates challenging samples to train a robust classifier for image classification.
Our approach outperforms the state of the art in various classification scenarios.
arXiv Detail & Related papers (2023-12-19T08:55:00Z) - Feature Weaken: Vicinal Data Augmentation for Classification [1.7013938542585925]
We use Feature Weaken to construct the vicinal data distribution with the same cosine similarity for model training.
This work can not only improve the classification performance and generalization of the model, but also stabilize the model training and accelerate the model convergence.
arXiv Detail & Related papers (2022-11-20T11:00:23Z) - Provably Learning Diverse Features in Multi-View Data with Midpoint Mixup [14.37428912254029]
Mixup is a data augmentation technique that relies on training using random convex combinations of data points and their labels.
We focus on classification problems in which each class may have multiple associated features (or views) that can be used to predict the class correctly.
Our main theoretical results demonstrate that, for a non-trivial class of data distributions with two features per class, training a 2-layer convolutional network using empirical risk minimization can lead to learning only one feature for almost all classes while training with a specific instantiation of Mixup succeeds in learning both features for every class.
arXiv Detail & Related papers (2022-10-24T18:11:37Z) - Learning with MISELBO: The Mixture Cookbook [62.75516608080322]
We present the first ever mixture of variational approximations for a normalizing flow-based hierarchical variational autoencoder (VAE) with VampPrior and a PixelCNN decoder network.
We explain this cooperative behavior by drawing a novel connection between VI and adaptive importance sampling.
We obtain state-of-the-art results among VAE architectures in terms of negative log-likelihood on the MNIST and FashionMNIST datasets.
arXiv Detail & Related papers (2022-09-30T15:01:35Z) - PRANC: Pseudo RAndom Networks for Compacting deep models [22.793523211040682]
PRANC enables significant compaction of a deep model.
In this study, we employ PRANC to condense image classification models and compress images by compacting their associated implicit neural networks.
arXiv Detail & Related papers (2022-06-16T22:03:35Z) - Boosting Discriminative Visual Representation Learning with
Scenario-Agnostic Mixup [54.09898347820941]
We propose textbfScenario-textbfAgnostic textbfMixup (SAMix) for both Self-supervised Learning (SSL) and supervised learning (SL) scenarios.
Specifically, we hypothesize and verify the objective function of mixup generation as optimizing local smoothness between two mixed classes.
A label-free generation sub-network is designed, which effectively provides non-trivial mixup samples and improves transferable abilities.
arXiv Detail & Related papers (2021-11-30T14:49:59Z) - Semi-Supervised Few-Shot Classification with Deep Invertible Hybrid
Models [4.189643331553922]
We propose a deep invertible hybrid model which integrates discriminative and generative learning at a latent space level for semi-supervised few-shot classification.
Our main originality lies in our integration of these components at a latent space level, which is effective in preventing overfitting.
arXiv Detail & Related papers (2021-05-22T05:55:16Z) - Jigsaw Clustering for Unsupervised Visual Representation Learning [68.09280490213399]
We propose a new jigsaw clustering pretext task in this paper.
Our method makes use of information from both intra- and inter-images.
It is even comparable to the contrastive learning methods when only half of training batches are used.
arXiv Detail & Related papers (2021-04-01T08:09:26Z) - ReMarNet: Conjoint Relation and Margin Learning for Small-Sample Image
Classification [49.87503122462432]
We introduce a novel neural network termed Relation-and-Margin learning Network (ReMarNet)
Our method assembles two networks of different backbones so as to learn the features that can perform excellently in both of the aforementioned two classification mechanisms.
Experiments on four image datasets demonstrate that our approach is effective in learning discriminative features from a small set of labeled samples.
arXiv Detail & Related papers (2020-06-27T13:50:20Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.