Gradient-Based Meta-Learning Using Uncertainty to Weigh Loss for
Few-Shot Learning
- URL: http://arxiv.org/abs/2208.08135v1
- Date: Wed, 17 Aug 2022 08:11:51 GMT
- Title: Gradient-Based Meta-Learning Using Uncertainty to Weigh Loss for
Few-Shot Learning
- Authors: Lin Ding, Peng Liu, Wenfeng Shen, Weijia Lu, Shengbo Chen
- Abstract summary: Model-Agnostic Meta-Learning (MAML) is one of the most successful meta-learning techniques for few-shot learning.
New method is proposed for task-specific learner adaptively learn to select parameters that minimize the loss of new tasks.
Method 1 generates weights by comparing meta-loss differences to improve the accuracy when there are few classes.
Method 2 introduces the homoscedastic uncertainty of each task to weigh multiple losses based on the original gradient descent.
- Score: 5.691930884128995
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Model-Agnostic Meta-Learning (MAML) is one of the most successful
meta-learning techniques for few-shot learning. It uses gradient descent to
learn commonalities between various tasks, enabling the model to learn the
meta-initialization of its own parameters to quickly adapt to new tasks using a
small amount of labeled training data. A key challenge to few-shot learning is
task uncertainty. Although a strong prior can be obtained from meta-learning
with a large number of tasks, a precision model of the new task cannot be
guaranteed because the volume of the training dataset is normally too small. In
this study, first,in the process of choosing initialization parameters, the new
method is proposed for task-specific learner adaptively learn to select
initialization parameters that minimize the loss of new tasks. Then, we propose
two improved methods for the meta-loss part: Method 1 generates weights by
comparing meta-loss differences to improve the accuracy when there are few
classes, and Method 2 introduces the homoscedastic uncertainty of each task to
weigh multiple losses based on the original gradient descent,as a way to
enhance the generalization ability to novel classes while ensuring accuracy
improvement. Compared with previous gradient-based meta-learning methods, our
model achieves better performance in regression tasks and few-shot
classification and improves the robustness of the model to the learning rate
and query sets in the meta-test set.
Related papers
- Learning to Learn with Indispensable Connections [6.040904021861969]
We propose a novel meta-learning method called Meta-LTH that includes indispensible (necessary) connections.
Our method improves the classification accuracy by approximately 2% (20-way 1-shot task setting) for omniglot dataset.
arXiv Detail & Related papers (2023-04-06T04:53:13Z) - Meta-free representation learning for few-shot learning via stochastic
weight averaging [13.6555672824229]
Recent studies on few-shot classification using transfer learning pose challenges to the effectiveness and efficiency of episodic meta-learning algorithms.
We propose a new transfer learning method to obtain accurate and reliable models for few-shot regression and classification.
arXiv Detail & Related papers (2022-04-26T17:36:34Z) - Large-Scale Meta-Learning with Continual Trajectory Shifting [76.29017270864308]
We show that allowing the meta-learners to take a larger number of inner gradient steps better captures the structure of heterogeneous and large-scale tasks.
In order to increase the frequency of meta-updates, we propose to estimate the required shift of the task-specific parameters.
We show that the algorithm largely outperforms the previous first-order meta-learning methods in terms of both generalization performance and convergence.
arXiv Detail & Related papers (2021-02-14T18:36:33Z) - Meta-Regularization by Enforcing Mutual-Exclusiveness [0.8057006406834467]
We propose a regularization technique for meta-learning models that gives the model designer more control over the information flow during meta-training.
Our proposed regularization function shows an accuracy boost of $sim$ $36%$ on the Omniglot dataset.
arXiv Detail & Related papers (2021-01-24T22:57:19Z) - Variable-Shot Adaptation for Online Meta-Learning [123.47725004094472]
We study the problem of learning new tasks from a small, fixed number of examples, by meta-learning across static data from a set of previous tasks.
We find that meta-learning solves the full task set with fewer overall labels and greater cumulative performance, compared to standard supervised methods.
These results suggest that meta-learning is an important ingredient for building learning systems that continuously learn and improve over a sequence of problems.
arXiv Detail & Related papers (2020-12-14T18:05:24Z) - A Primal-Dual Subgradient Approachfor Fair Meta Learning [23.65344558042896]
Few shot meta-learning is well-known with its fast-adapted capability and accuracy generalization onto unseen tasks.
We propose a Primal-Dual Fair Meta-learning framework, namely PDFM, which learns to train fair machine learning models using only a few examples.
arXiv Detail & Related papers (2020-09-26T19:47:38Z) - Few Is Enough: Task-Augmented Active Meta-Learning for Brain Cell
Classification [8.998976678920236]
We propose a tAsk-auGmented actIve meta-LEarning (AGILE) method to efficiently adapt Deep Neural Networks to new tasks.
AGILE combines a meta-learning algorithm with a novel task augmentation technique which we use to generate an initial adaptive model.
We show that the proposed task-augmented meta-learning framework can learn to classify new cell types after a single gradient step.
arXiv Detail & Related papers (2020-07-09T18:03:12Z) - Incremental Object Detection via Meta-Learning [77.55310507917012]
We propose a meta-learning approach that learns to reshape model gradients, such that information across incremental tasks is optimally shared.
In comparison to existing meta-learning methods, our approach is task-agnostic, allows incremental addition of new-classes and scales to high-capacity models for object detection.
arXiv Detail & Related papers (2020-03-17T13:40:00Z) - Meta-Baseline: Exploring Simple Meta-Learning for Few-Shot Learning [79.25478727351604]
We explore a simple process: meta-learning over a whole-classification pre-trained model on its evaluation metric.
We observe this simple method achieves competitive performance to state-of-the-art methods on standard benchmarks.
arXiv Detail & Related papers (2020-03-09T20:06:36Z) - Meta Cyclical Annealing Schedule: A Simple Approach to Avoiding
Meta-Amortization Error [50.83356836818667]
We develop a novel meta-regularization objective using it cyclical annealing schedule and it maximum mean discrepancy (MMD) criterion.
The experimental results show that our approach substantially outperforms standard meta-learning algorithms.
arXiv Detail & Related papers (2020-03-04T04:43:16Z) - Incremental Meta-Learning via Indirect Discriminant Alignment [118.61152684795178]
We develop a notion of incremental learning during the meta-training phase of meta-learning.
Our approach performs favorably at test time as compared to training a model with the full meta-training set.
arXiv Detail & Related papers (2020-02-11T01:39:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.