Generating meta-learning tasks to evolve parametric loss for
classification learning
- URL: http://arxiv.org/abs/2111.10583v1
- Date: Sat, 20 Nov 2021 13:07:55 GMT
- Title: Generating meta-learning tasks to evolve parametric loss for
classification learning
- Authors: Zhaoyang Hai, Xiabi Liu, Yuchen Ren, Nouman Q. Soomro
- Abstract summary: In existing meta-learning approaches, learning tasks for training meta-models are usually collected from public datasets.
We propose a meta-learning approach based on randomly generated meta-learning tasks to obtain a parametric loss for classification learning based on big data.
- Score: 1.1355370218310157
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The field of meta-learning has seen a dramatic rise in interest in recent
years. In existing meta-learning approaches, learning tasks for training
meta-models are usually collected from public datasets, which brings the
difficulty of obtaining a sufficient number of meta-learning tasks with a large
amount of training data. In this paper, we propose a meta-learning approach
based on randomly generated meta-learning tasks to obtain a parametric loss for
classification learning based on big data. The loss is represented by a deep
neural network, called meta-loss network (MLN). To train the MLN, we construct
a large number of classification learning tasks through randomly generating
training data, validation data, and corresponding ground-truth linear
classifier. Our approach has two advantages. First, sufficient meta-learning
tasks with large number of training data can be obtained easily. Second, the
ground-truth classifier is given, so that the difference between the learned
classifier and the ground-truth model can be measured to reflect the
performance of MLN more precisely than validation accuracy. Based on this
difference, we apply the evolutionary strategy algorithm to find out the
optimal MLN. The resultant MLN not only leads to satisfactory learning effects
on generated linear classifier learning tasks for testing, but also behaves
very well on generated nonlinear classifier learning tasks and various public
classification tasks. Our MLN stably surpass cross-entropy (CE) and mean square
error (MSE) in testing accuracy and generalization ability. These results
illustrate the possibility of achieving satisfactory meta-learning effects
using generated learning tasks.
Related papers
- Task-Distributionally Robust Data-Free Meta-Learning [99.56612787882334]
Data-Free Meta-Learning (DFML) aims to efficiently learn new tasks by leveraging multiple pre-trained models without requiring their original training data.
For the first time, we reveal two major challenges hindering their practical deployments: Task-Distribution Shift ( TDS) and Task-Distribution Corruption (TDC)
arXiv Detail & Related papers (2023-11-23T15:46:54Z) - Unsupervised Representation Learning to Aid Semi-Supervised Meta
Learning [16.534014215010757]
We propose a one-shot unsupervised meta-learning to learn latent representation of training samples.
A temperature-scaled cross-entropy loss is used in the inner loop of meta-learning to prevent overfitting.
The proposed method is model agnostic and can aid any meta-learning model to improve accuracy.
arXiv Detail & Related papers (2023-10-19T18:25:22Z) - Understanding Transfer Learning and Gradient-Based Meta-Learning
Techniques [5.2997197698288945]
We investigate performance differences between fine, MAML, and another meta-learning technique called Reptile.
Our findings show that both the output layer and the noisy training conditions induced by data scarcity play important roles in facilitating this specialization for MAML.
We show that the pre-trained features as obtained by the finetuning baseline are more diverse and discriminative than those learned by MAML and Reptile.
arXiv Detail & Related papers (2023-10-09T20:51:49Z) - Theoretical Characterization of the Generalization Performance of
Overfitted Meta-Learning [70.52689048213398]
This paper studies the performance of overfitted meta-learning under a linear regression model with Gaussian features.
We find new and interesting properties that do not exist in single-task linear regression.
Our analysis suggests that benign overfitting is more significant and easier to observe when the noise and the diversity/fluctuation of the ground truth of each training task are large.
arXiv Detail & Related papers (2023-04-09T20:36:13Z) - Learning to Learn with Indispensable Connections [6.040904021861969]
We propose a novel meta-learning method called Meta-LTH that includes indispensible (necessary) connections.
Our method improves the classification accuracy by approximately 2% (20-way 1-shot task setting) for omniglot dataset.
arXiv Detail & Related papers (2023-04-06T04:53:13Z) - Meta-Learning with Self-Improving Momentum Target [72.98879709228981]
We propose Self-improving Momentum Target (SiMT) to improve the performance of a meta-learner.
SiMT generates the target model by adapting from the temporal ensemble of the meta-learner.
We show that SiMT brings a significant performance gain when combined with a wide range of meta-learning methods.
arXiv Detail & Related papers (2022-10-11T06:45:15Z) - ProtoDA: Efficient Transfer Learning for Few-Shot Intent Classification [21.933876113300897]
We adopt an alternative approach by transfer learning on an ensemble of related tasks using prototypical networks under the meta-learning paradigm.
Using intent classification as a case study, we demonstrate that increasing variability in training tasks can significantly improve classification performance.
arXiv Detail & Related papers (2021-01-28T00:19:13Z) - Expert Training: Task Hardness Aware Meta-Learning for Few-Shot
Classification [62.10696018098057]
We propose an easy-to-hard expert meta-training strategy to arrange the training tasks properly.
A task hardness aware module is designed and integrated into the training procedure to estimate the hardness of a task.
Experimental results on the miniImageNet and tieredImageNetSketch datasets show that the meta-learners can obtain better results with our expert training strategy.
arXiv Detail & Related papers (2020-07-13T08:49:00Z) - Few Is Enough: Task-Augmented Active Meta-Learning for Brain Cell
Classification [8.998976678920236]
We propose a tAsk-auGmented actIve meta-LEarning (AGILE) method to efficiently adapt Deep Neural Networks to new tasks.
AGILE combines a meta-learning algorithm with a novel task augmentation technique which we use to generate an initial adaptive model.
We show that the proposed task-augmented meta-learning framework can learn to classify new cell types after a single gradient step.
arXiv Detail & Related papers (2020-07-09T18:03:12Z) - Provable Meta-Learning of Linear Representations [114.656572506859]
We provide fast, sample-efficient algorithms to address the dual challenges of learning a common set of features from multiple, related tasks, and transferring this knowledge to new, unseen tasks.
We also provide information-theoretic lower bounds on the sample complexity of learning these linear features.
arXiv Detail & Related papers (2020-02-26T18:21:34Z) - Incremental Meta-Learning via Indirect Discriminant Alignment [118.61152684795178]
We develop a notion of incremental learning during the meta-training phase of meta-learning.
Our approach performs favorably at test time as compared to training a model with the full meta-training set.
arXiv Detail & Related papers (2020-02-11T01:39:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.