Navigating the Trade-Off between Multi-Task Learning and Learning to
Multitask in Deep Neural Networks
- URL: http://arxiv.org/abs/2007.10527v2
- Date: Tue, 5 Jan 2021 18:16:35 GMT
- Title: Navigating the Trade-Off between Multi-Task Learning and Learning to
Multitask in Deep Neural Networks
- Authors: Sachin Ravi and Sebastian Musslick and Maia Hamin and Theodore L.
Willke and Jonathan D. Cohen
- Abstract summary: Multi-task learning refers to a paradigm in machine learning in which a network is trained on various related tasks to facilitate the acquisition of tasks.
multitasking is used to indicate, especially in the cognitive science literature, the ability to execute multiple tasks simultaneously.
We show that the same tension arises in deep networks and discuss a meta-learning algorithm for an agent to manage this trade-off in an unfamiliar environment.
- Score: 9.278739724750343
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The terms multi-task learning and multitasking are easily confused.
Multi-task learning refers to a paradigm in machine learning in which a network
is trained on various related tasks to facilitate the acquisition of tasks. In
contrast, multitasking is used to indicate, especially in the cognitive science
literature, the ability to execute multiple tasks simultaneously. While
multi-task learning exploits the discovery of common structure between tasks in
the form of shared representations, multitasking is promoted by separating
representations between tasks to avoid processing interference. Here, we build
on previous work involving shallow networks and simple task settings suggesting
that there is a trade-off between multi-task learning and multitasking,
mediated by the use of shared versus separated representations. We show that
the same tension arises in deep networks and discuss a meta-learning algorithm
for an agent to manage this trade-off in an unfamiliar environment. We display
through different experiments that the agent is able to successfully optimize
its training strategy as a function of the environment.
Related papers
- Saliency-Regularized Deep Multi-Task Learning [7.3810864598379755]
Multitask learning enforces multiple learning tasks to share knowledge to improve their generalization abilities.
Modern deep multitask learning can jointly learn latent features and task sharing, but they are obscure in task relation.
This paper proposes a new multitask learning framework that jointly learns latent features and explicit task relations.
arXiv Detail & Related papers (2022-07-03T20:26:44Z) - Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners [67.5865966762559]
We study whether sparsely activated Mixture-of-Experts (MoE) improve multi-task learning.
We devise task-aware gating functions to route examples from different tasks to specialized experts.
This results in a sparsely activated multi-task model with a large number of parameters, but with the same computational cost as that of a dense model.
arXiv Detail & Related papers (2022-04-16T00:56:12Z) - Modular Adaptive Policy Selection for Multi-Task Imitation Learning
through Task Division [60.232542918414985]
Multi-task learning often suffers from negative transfer, sharing information that should be task-specific.
This is done by using proto-policies as modules to divide the tasks into simple sub-behaviours that can be shared.
We also demonstrate its ability to autonomously divide the tasks into both shared and task-specific sub-behaviours.
arXiv Detail & Related papers (2022-03-28T15:53:17Z) - Multi-Task Neural Processes [105.22406384964144]
We develop multi-task neural processes, a new variant of neural processes for multi-task learning.
In particular, we propose to explore transferable knowledge from related tasks in the function space to provide inductive bias for improving each individual task.
Results demonstrate the effectiveness of multi-task neural processes in transferring useful knowledge among tasks for multi-task learning.
arXiv Detail & Related papers (2021-11-10T17:27:46Z) - On the relationship between disentanglement and multi-task learning [62.997667081978825]
We take a closer look at the relationship between disentanglement and multi-task learning based on hard parameter sharing.
We show that disentanglement appears naturally during the process of multi-task neural network training.
arXiv Detail & Related papers (2021-10-07T14:35:34Z) - Multi-Task Learning with Sequence-Conditioned Transporter Networks [67.57293592529517]
We aim to solve multi-task learning through the lens of sequence-conditioning and weighted sampling.
We propose a new suite of benchmark aimed at compositional tasks, MultiRavens, which allows defining custom task combinations.
Second, we propose a vision-based end-to-end system architecture, Sequence-Conditioned Transporter Networks, which augments Goal-Conditioned Transporter Networks with sequence-conditioning and weighted sampling.
arXiv Detail & Related papers (2021-09-15T21:19:11Z) - MultiTask-CenterNet (MCN): Efficient and Diverse Multitask Learning
using an Anchor Free Approach [0.13764085113103217]
Multitask learning is a common approach in machine learning.
In this paper we augment the CenterNet anchor-free approach for training multiple perception related tasks together.
arXiv Detail & Related papers (2021-08-11T06:57:04Z) - Multitask learning over graphs: An Approach for Distributed, Streaming
Machine Learning [46.613346075513206]
Multitask learning is an approach to inductive transfer learning.
Recent years have witnessed an increasing ability to collect data in a distributed and streaming manner.
This requires the design of new strategies for learning jointly multiple tasks from streaming data over distributed (or networked) systems.
arXiv Detail & Related papers (2020-01-07T15:32:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.