Attentive Feature Reuse for Multi Task Meta learning
- URL: http://arxiv.org/abs/2006.07438v1
- Date: Fri, 12 Jun 2020 19:33:11 GMT
- Title: Attentive Feature Reuse for Multi Task Meta learning
- Authors: Kiran Lekkala, Laurent Itti
- Abstract summary: We develop new algorithms for simultaneous learning of multiple tasks.
We propose an attention mechanism to dynamically specialize the network, at runtime, for each task.
Our method improves performance on new, previously unseen environments.
- Score: 17.8055398673228
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We develop new algorithms for simultaneous learning of multiple tasks (e.g.,
image classification, depth estimation), and for adapting to unseen task/domain
distributions within those high-level tasks (e.g., different environments).
First, we learn common representations underlying all tasks. We then propose an
attention mechanism to dynamically specialize the network, at runtime, for each
task. Our approach is based on weighting each feature map of the backbone
network, based on its relevance to a particular task. To achieve this, we
enable the attention module to learn task representations during training,
which are used to obtain attention weights. Our method improves performance on
new, previously unseen environments, and is 1.5x faster than standard existing
meta learning methods using similar architectures. We highlight performance
improvements for Multi-Task Meta Learning of 4 tasks (image classification,
depth, vanishing point, and surface normal estimation), each over 10 to 25 test
domains/environments, a result that could not be achieved with standard meta
learning techniques like MAML.
Related papers
- Meta-Learning with Heterogeneous Tasks [42.695853959923625]
Heterogeneous Tasks Robust Meta-learning (HeTRoM)
An efficient iterative optimization algorithm based on bi-level optimization.
Results demonstrate that our method provides flexibility, enabling users to adapt to diverse task settings.
arXiv Detail & Related papers (2024-10-24T16:32:23Z) - YOLOR-Based Multi-Task Learning [12.5920336941241]
Multi-task learning (MTL) aims to learn multiple tasks using a single model and jointly improve all of them assuming generalization and shared semantics.
We propose building on You Only Learn One Representation (YOLOR), a network architecture specifically designed for multitasking.
We find that our method achieves competitive performance on all tasks while maintaining a low parameter count and without any pre-training.
arXiv Detail & Related papers (2023-09-29T01:42:21Z) - Multi-Task Meta Learning: learn how to adapt to unseen tasks [4.287114092271669]
This work proposes Multi-task Meta Learning (MTML), integrating two learning paradigms Multi-Task Learning (MTL) and meta learning.
The fundamental idea is to train a multi-task model, such that when an unseen task is introduced, it can learn in fewer steps whilst offering a performance at least as good as conventional single task learning.
MTML achieves state-of-the-art results for three out of four tasks for the NYU-v2 dataset and two out of four for the taskonomy dataset.
arXiv Detail & Related papers (2022-10-13T12:59:54Z) - Fast Inference and Transfer of Compositional Task Structures for
Few-shot Task Generalization [101.72755769194677]
We formulate it as a few-shot reinforcement learning problem where a task is characterized by a subtask graph.
Our multi-task subtask graph inferencer (MTSGI) first infers the common high-level task structure in terms of the subtask graph from the training tasks.
Our experiment results on 2D grid-world and complex web navigation domains show that the proposed method can learn and leverage the common underlying structure of the tasks for faster adaptation to the unseen tasks.
arXiv Detail & Related papers (2022-05-25T10:44:25Z) - Set-based Meta-Interpolation for Few-Task Meta-Learning [79.4236527774689]
We propose a novel domain-agnostic task augmentation method, Meta-Interpolation, to densify the meta-training task distribution.
We empirically validate the efficacy of Meta-Interpolation on eight datasets spanning across various domains.
arXiv Detail & Related papers (2022-05-20T06:53:03Z) - Meta-Learning with Fewer Tasks through Task Interpolation [67.03769747726666]
Current meta-learning algorithms require a large number of meta-training tasks, which may not be accessible in real-world scenarios.
By meta-learning with task gradient (MLTI), our approach effectively generates additional tasks by randomly sampling a pair of tasks and interpolating the corresponding features and labels.
Empirically, in our experiments on eight datasets from diverse domains, we find that the proposed general MLTI framework is compatible with representative meta-learning algorithms and consistently outperforms other state-of-the-art strategies.
arXiv Detail & Related papers (2021-06-04T20:15:34Z) - Adaptive Task Sampling for Meta-Learning [79.61146834134459]
Key idea of meta-learning for few-shot classification is to mimic the few-shot situations faced at test time.
We propose an adaptive task sampling method to improve the generalization performance.
arXiv Detail & Related papers (2020-07-17T03:15:53Z) - MTL-NAS: Task-Agnostic Neural Architecture Search towards
General-Purpose Multi-Task Learning [71.90902837008278]
We propose to incorporate neural architecture search (NAS) into general-purpose multi-task learning (GP-MTL)
In order to adapt to different task combinations, we disentangle the GP-MTL networks into single-task backbones.
We also propose a novel single-shot gradient-based search algorithm that closes the performance gap between the searched architectures.
arXiv Detail & Related papers (2020-03-31T09:49:14Z) - Adversarial Continual Learning [99.56738010842301]
We propose a hybrid continual learning framework that learns a disjoint representation for task-invariant and task-specific features.
Our model combines architecture growth to prevent forgetting of task-specific skills and an experience replay approach to preserve shared skills.
arXiv Detail & Related papers (2020-03-21T02:08:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.