The Curse of Low Task Diversity: On the Failure of Transfer Learning to
Outperform MAML and Their Empirical Equivalence
- URL: http://arxiv.org/abs/2208.01545v1
- Date: Tue, 2 Aug 2022 15:49:11 GMT
- Title: The Curse of Low Task Diversity: On the Failure of Transfer Learning to
Outperform MAML and Their Empirical Equivalence
- Authors: Brando Miranda, Patrick Yu, Yu-Xiong Wang, Sanmi Koyejo
- Abstract summary: We propose a novel metric -- the diversity coefficient -- to measure the diversity of tasks in a few-shot learning benchmark.
Using the diversity coefficient, we show that the popular MiniImageNet and CIFAR-FS few-shot learning benchmarks have low diversity.
- Score: 20.965759895300327
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, it has been observed that a transfer learning solution might be all
we need to solve many few-shot learning benchmarks -- thus raising important
questions about when and how meta-learning algorithms should be deployed. In
this paper, we seek to clarify these questions by 1. proposing a novel metric
-- the diversity coefficient -- to measure the diversity of tasks in a few-shot
learning benchmark and 2. by comparing Model-Agnostic Meta-Learning (MAML) and
transfer learning under fair conditions (same architecture, same optimizer, and
all models trained to convergence). Using the diversity coefficient, we show
that the popular MiniImageNet and CIFAR-FS few-shot learning benchmarks have
low diversity. This novel insight contextualizes claims that transfer learning
solutions are better than meta-learned solutions in the regime of low diversity
under a fair comparison. Specifically, we empirically find that a low diversity
coefficient correlates with a high similarity between transfer learning and
MAML learned solutions in terms of accuracy at meta-test time and
classification layer similarity (using feature based distance metrics like
SVCCA, PWCCA, CKA, and OPD). To further support our claim, we find this
meta-test accuracy holds even as the model size changes. Therefore, we conclude
that in the low diversity regime, MAML and transfer learning have equivalent
meta-test performance when both are compared fairly. We also hope our work
inspires more thoughtful constructions and quantitative evaluations of
meta-learning benchmarks in the future.
Related papers
- Learning to Learn with Indispensable Connections [6.040904021861969]
We propose a novel meta-learning method called Meta-LTH that includes indispensible (necessary) connections.
Our method improves the classification accuracy by approximately 2% (20-way 1-shot task setting) for omniglot dataset.
arXiv Detail & Related papers (2023-04-06T04:53:13Z) - NEVIS'22: A Stream of 100 Tasks Sampled from 30 Years of Computer Vision
Research [96.53307645791179]
We introduce the Never-Ending VIsual-classification Stream (NEVIS'22), a benchmark consisting of a stream of over 100 visual classification tasks.
Despite being limited to classification, the resulting stream has a rich diversity of tasks from OCR, to texture analysis, scene recognition, and so forth.
Overall, NEVIS'22 poses an unprecedented challenge for current sequential learning approaches due to the scale and diversity of tasks.
arXiv Detail & Related papers (2022-11-15T18:57:46Z) - Contrastive Knowledge-Augmented Meta-Learning for Few-Shot
Classification [28.38744876121834]
We introduce CAML (Contrastive Knowledge-Augmented Meta Learning), a novel approach for knowledge-enhanced few-shot learning.
We evaluate the performance of CAML in different few-shot learning scenarios.
arXiv Detail & Related papers (2022-07-25T17:01:29Z) - Learning What Not to Segment: A New Perspective on Few-Shot Segmentation [63.910211095033596]
Recently few-shot segmentation (FSS) has been extensively developed.
This paper proposes a fresh and straightforward insight to alleviate the problem.
In light of the unique nature of the proposed approach, we also extend it to a more realistic but challenging setting.
arXiv Detail & Related papers (2022-03-15T03:08:27Z) - The Effect of Diversity in Meta-Learning [79.56118674435844]
Few-shot learning aims to learn representations that can tackle novel tasks given a small number of examples.
Recent studies show that task distribution plays a vital role in the model's performance.
We study different task distributions on a myriad of models and datasets to evaluate the effect of task diversity on meta-learning algorithms.
arXiv Detail & Related papers (2022-01-27T19:39:07Z) - The Curse of Zero Task Diversity: On the Failure of Transfer Learning to
Outperform MAML and their Empirical Equivalence [19.556093984142418]
A transfer learning solution might be all we needed to solve many few-shot learning benchmarks.
We name this metric the diversity coefficient of a few-shot learning benchmark.
We show that when making a fair comparison between MAML learned solutions to transfer learning, both have identical meta-test accuracy.
arXiv Detail & Related papers (2021-12-24T18:42:58Z) - Comparing Transfer and Meta Learning Approaches on a Unified Few-Shot
Classification Benchmark [44.530605715850506]
Cross-family study of the best transfer and meta learners on a large-scale meta-learning benchmark and a transfer learning benchmark.
We find that, on average, large-scale transfer methods (Big Transfer, BiT) outperform competing approaches on MD, even when trained only on ImageNet.
We reveal a number of discrepancies in evaluation norms and study some of these in light of the performance gap.
arXiv Detail & Related papers (2021-04-06T16:17:51Z) - A Taxonomy of Similarity Metrics for Markov Decision Processes [62.997667081978825]
In recent years, transfer learning has succeeded in making Reinforcement Learning (RL) algorithms more efficient.
In this paper, we propose a categorization of these metrics and analyze the definitions of similarity proposed so far.
arXiv Detail & Related papers (2021-03-08T12:36:42Z) - Revisiting Unsupervised Meta-Learning: Amplifying or Compensating for
the Characteristics of Few-Shot Tasks [30.893785366366078]
We develop a practical approach towards few-shot image classification, where a visual recognition system is constructed with limited data.
We find that the base class set labels are not necessary, and discriminative embeddings could be meta-learned in an unsupervised manner.
Experiments on few-shot learning benchmarks verify our approaches outperform previous methods by a 4-10% performance gap.
arXiv Detail & Related papers (2020-11-30T10:08:35Z) - Meta-Baseline: Exploring Simple Meta-Learning for Few-Shot Learning [79.25478727351604]
We explore a simple process: meta-learning over a whole-classification pre-trained model on its evaluation metric.
We observe this simple method achieves competitive performance to state-of-the-art methods on standard benchmarks.
arXiv Detail & Related papers (2020-03-09T20:06:36Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.