Meta-learning the Learning Trends Shared Across Tasks
- URL: http://arxiv.org/abs/2010.09291v1
- Date: Mon, 19 Oct 2020 08:06:47 GMT
- Title: Meta-learning the Learning Trends Shared Across Tasks
- Authors: Jathushan Rajasegaran, Salman Khan, Munawar Hayat, Fahad Shahbaz Khan,
Mubarak Shah
- Abstract summary: Gradient-based meta-learning algorithms excel at quick adaptation to new tasks with limited data.
Existing meta-learning approaches only depend on the current task information during the adaptation.
We propose a 'Path-aware' model-agnostic meta-learning approach.
- Score: 123.10294801296926
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Meta-learning stands for 'learning to learn' such that generalization to new
tasks is achieved. Among these methods, Gradient-based meta-learning algorithms
are a specific sub-class that excel at quick adaptation to new tasks with
limited data. This demonstrates their ability to acquire transferable
knowledge, a capability that is central to human learning. However, the
existing meta-learning approaches only depend on the current task information
during the adaptation, and do not share the meta-knowledge of how a similar
task has been adapted before. To address this gap, we propose a 'Path-aware'
model-agnostic meta-learning approach. Specifically, our approach not only
learns a good initialization for adaptation, it also learns an optimal way to
adapt these parameters to a set of task-specific parameters, with learnable
update directions, learning rates and, most importantly, the way updates evolve
over different time-steps. Compared to the existing meta-learning methods, our
approach offers: (a) The ability to learn gradient-preconditioning at different
time-steps of the inner-loop, thereby modeling the dynamic learning behavior
shared across tasks, and (b) The capability of aggregating the learning context
through the provision of direct gradient-skip connections from the old
time-steps, thus avoiding overfitting and improving generalization. In essence,
our approach not only learns a transferable initialization, but also models the
optimal update directions, learning rates, and task-specific learning trends.
Specifically, in terms of learning trends, our approach determines the way
update directions shape up as the task-specific learning progresses and how the
previous update history helps in the current update. Our approach is simple to
implement and demonstrates faster convergence. We report significant
performance improvements on a number of FSL datasets.
Related papers
- Learning to Learn without Forgetting using Attention [5.6739565497512405]
Continual learning (CL) refers to the ability to continually learn over time by accommodating new knowledge while retaining previously learned experience.
Current machine learning methods are highly prone to overwrite previously learned patterns and thus forget past experience.
Since hand-crafting effective update mechanisms is difficult, we propose meta-learning a transformer-based to enhance CL.
arXiv Detail & Related papers (2024-08-06T14:25:23Z) - Meta-Learning and representation learner: A short theoretical note [0.0]
Meta-learning is a subfield of machine learning where the goal is to develop models and algorithms that can learn from various tasks.
Unlike traditional machine learning methods focusing on learning a specific task, meta-learning aims to leverage experience from previous tasks to enhance future learning.
arXiv Detail & Related papers (2024-07-04T23:47:10Z) - Algorithm Design for Online Meta-Learning with Task Boundary Detection [63.284263611646]
We propose a novel algorithm for task-agnostic online meta-learning in non-stationary environments.
We first propose two simple but effective detection mechanisms of task switches and distribution shift.
We show that a sublinear task-averaged regret can be achieved for our algorithm under mild conditions.
arXiv Detail & Related papers (2023-02-02T04:02:49Z) - On the Effectiveness of Fine-tuning Versus Meta-reinforcement Learning [71.55412580325743]
We show that multi-task pretraining with fine-tuning on new tasks performs equally as well, or better, than meta-pretraining with meta test-time adaptation.
This is encouraging for future research, as multi-task pretraining tends to be simpler and computationally cheaper than meta-RL.
arXiv Detail & Related papers (2022-06-07T13:24:00Z) - Dynamic Regret Analysis for Online Meta-Learning [0.0]
The online meta-learning framework has arisen as a powerful tool for the continual lifelong learning setting.
This formulation involves two levels: outer level which learns meta-learners and inner level which learns task-specific models.
We establish performance in terms of dynamic regret which handles changing environments from a global prospective.
We carry out our analyses in a setting, and in expectation prove a logarithmic local dynamic regret which explicitly depends on the total number of iterations.
arXiv Detail & Related papers (2021-09-29T12:12:59Z) - META-Learning Eligibility Traces for More Sample Efficient Temporal
Difference Learning [2.0559497209595823]
We propose a meta-learning method for adjusting the eligibility trace parameter, in a state-dependent manner.
The adaptation is achieved with the help of auxiliary learners that learn distributional information about the update targets online.
We prove that, under some assumptions, the proposed method improves the overall quality of the update targets, by minimizing the overall target error.
arXiv Detail & Related papers (2020-06-16T03:41:07Z) - Meta-Reinforcement Learning Robust to Distributional Shift via Model
Identification and Experience Relabeling [126.69933134648541]
We present a meta-reinforcement learning algorithm that is both efficient and extrapolates well when faced with out-of-distribution tasks at test time.
Our method is based on a simple insight: we recognize that dynamics models can be adapted efficiently and consistently with off-policy data.
arXiv Detail & Related papers (2020-06-12T13:34:46Z) - Incremental Object Detection via Meta-Learning [77.55310507917012]
We propose a meta-learning approach that learns to reshape model gradients, such that information across incremental tasks is optimally shared.
In comparison to existing meta-learning methods, our approach is task-agnostic, allows incremental addition of new-classes and scales to high-capacity models for object detection.
arXiv Detail & Related papers (2020-03-17T13:40:00Z) - Revisiting Meta-Learning as Supervised Learning [69.2067288158133]
We aim to provide a principled, unifying framework by revisiting and strengthening the connection between meta-learning and traditional supervised learning.
By treating pairs of task-specific data sets and target models as (feature, label) samples, we can reduce many meta-learning algorithms to instances of supervised learning.
This view not only unifies meta-learning into an intuitive and practical framework but also allows us to transfer insights from supervised learning directly to improve meta-learning.
arXiv Detail & Related papers (2020-02-03T06:13:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.