Transductive Information Maximization For Few-Shot Learning
- URL: http://arxiv.org/abs/2008.11297v3
- Date: Fri, 23 Oct 2020 17:36:19 GMT
- Title: Transductive Information Maximization For Few-Shot Learning
- Authors: Malik Boudiaf, Ziko Imtiaz Masud, J\'er\^ome Rony, Jos\'e Dolz, Pablo
Piantanida, Ismail Ben Ayed
- Abstract summary: We introduce Transductive Infomation Maximization (TIM) for few-shot learning.
Our method maximizes the mutual information between the query features and their label predictions for a given few-shot task.
We propose a new alternating-direction solver for our mutual-information loss.
- Score: 41.461586994394565
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce Transductive Infomation Maximization (TIM) for few-shot
learning. Our method maximizes the mutual information between the query
features and their label predictions for a given few-shot task, in conjunction
with a supervision loss based on the support set. Furthermore, we propose a new
alternating-direction solver for our mutual-information loss, which
substantially speeds up transductive-inference convergence over gradient-based
optimization, while yielding similar accuracy. TIM inference is modular: it can
be used on top of any base-training feature extractor. Following standard
transductive few-shot settings, our comprehensive experiments demonstrate that
TIM outperforms state-of-the-art methods significantly across various datasets
and networks, while used on top of a fixed feature extractor trained with
simple cross-entropy on the base classes, without resorting to complex
meta-learning schemes. It consistently brings between 2% and 5% improvement in
accuracy over the best performing method, not only on all the well-established
few-shot benchmarks but also on more challenging scenarios,with domain shifts
and larger numbers of classes.
Related papers
- Revisiting Consistency Regularization for Semi-Supervised Learning [80.28461584135967]
We propose an improved consistency regularization framework by a simple yet effective technique, FeatDistLoss.
Experimental results show that our model defines a new state of the art for various datasets and settings.
arXiv Detail & Related papers (2021-12-10T20:46:13Z) - Squeezing Backbone Feature Distributions to the Max for Efficient
Few-Shot Learning [3.1153758106426603]
Few-shot classification is a challenging problem due to the uncertainty caused by using few labelled samples.
We propose a novel transfer-based method which aims at processing the feature vectors so that they become closer to Gaussian-like distributions.
In the case of transductive few-shot learning where unlabelled test samples are available during training, we also introduce an optimal-transport inspired algorithm to boost even further the achieved performance.
arXiv Detail & Related papers (2021-10-18T16:29:17Z) - Mutual-Information Based Few-Shot Classification [34.95314059362982]
We introduce Transductive Infomation Maximization (TIM) for few-shot learning.
Our method maximizes the mutual information between the query features and their label predictions for a given few-shot task.
We propose a new alternating-direction solver, which speeds up transductive inference over gradient-based optimization.
arXiv Detail & Related papers (2021-06-23T09:17:23Z) - Towards Accurate Knowledge Transfer via Target-awareness Representation
Disentanglement [56.40587594647692]
We propose a novel transfer learning algorithm, introducing the idea of Target-awareness REpresentation Disentanglement (TRED)
TRED disentangles the relevant knowledge with respect to the target task from the original source model and used as a regularizer during fine-tuning the target model.
Experiments on various real world datasets show that our method stably improves the standard fine-tuning by more than 2% in average.
arXiv Detail & Related papers (2020-10-16T17:45:08Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z) - Laplacian Regularized Few-Shot Learning [35.381119443377195]
We propose a transductive Laplacian-regularized inference for few-shot tasks.
Our inference does not re-train the base model, and can be viewed as a graph clustering of the query set.
Our LaplacianShot consistently outperforms state-of-the-art methods by significant margins across different models.
arXiv Detail & Related papers (2020-06-28T02:17:52Z) - A Transductive Multi-Head Model for Cross-Domain Few-Shot Learning [72.30054522048553]
We present a new method, Transductive Multi-Head Few-Shot learning (TMHFS), to address the Cross-Domain Few-Shot Learning challenge.
The proposed methods greatly outperform the strong baseline, fine-tuning, on four different target domains.
arXiv Detail & Related papers (2020-06-08T02:39:59Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.