DynaShare: Task and Instance Conditioned Parameter Sharing for
Multi-Task Learning
- URL: http://arxiv.org/abs/2305.17305v1
- Date: Fri, 26 May 2023 23:43:21 GMT
- Title: DynaShare: Task and Instance Conditioned Parameter Sharing for
Multi-Task Learning
- Authors: Elahe Rahimian, Golara Javadi, Frederick Tung, Gabriel Oliveira
- Abstract summary: We present a novel parameter sharing method for multi-task learning.
We propose to dynamically decide which parts of the network to activate based on both the task and the input instance.
Our approach learns a hierarchical gating policy consisting of a task-specific policy for coarse layer selection and gating units for individual input instances.
- Score: 11.955637263520492
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multi-task networks rely on effective parameter sharing to achieve robust
generalization across tasks. In this paper, we present a novel parameter
sharing method for multi-task learning that conditions parameter sharing on
both the task and the intermediate feature representations at inference time.
In contrast to traditional parameter sharing approaches, which fix or learn a
deterministic sharing pattern during training and apply the same pattern to all
examples during inference, we propose to dynamically decide which parts of the
network to activate based on both the task and the input instance. Our approach
learns a hierarchical gating policy consisting of a task-specific policy for
coarse layer selection and gating units for individual input instances, which
work together to determine the execution path at inference time. Experiments on
the NYU v2, Cityscapes and MIMIC-III datasets demonstrate the potential of the
proposed approach and its applicability across problem domains.
Related papers
- Task-Aware Harmony Multi-Task Decision Transformer for Offline Reinforcement Learning [70.96345405979179]
The purpose of offline multi-task reinforcement learning (MTRL) is to develop a unified policy applicable to diverse tasks without the need for online environmental interaction.
variations in task content and complexity pose significant challenges in policy formulation.
We introduce the Harmony Multi-Task Decision Transformer (HarmoDT), a novel solution designed to identify an optimal harmony subspace of parameters for each task.
arXiv Detail & Related papers (2024-11-02T05:49:14Z) - RepVF: A Unified Vector Fields Representation for Multi-task 3D Perception [64.80760846124858]
This paper proposes a novel unified representation, RepVF, which harmonizes the representation of various perception tasks.
RepVF characterizes the structure of different targets in the scene through a vector field, enabling a single-head, multi-task learning model.
Building upon RepVF, we introduce RFTR, a network designed to exploit the inherent connections between different tasks.
arXiv Detail & Related papers (2024-07-15T16:25:07Z) - PaCo: Parameter-Compositional Multi-Task Reinforcement Learning [44.43196786555784]
We introduce a parameter-compositional approach (PaCo) as an attempt to address these challenges.
Policies for all the single tasks lie in this subspace and can be composed by interpolating with the learned set.
We demonstrate the state-of-the-art performance on Meta-World benchmarks, verifying the effectiveness of the proposed approach.
arXiv Detail & Related papers (2022-10-21T01:00:10Z) - Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners [67.5865966762559]
We study whether sparsely activated Mixture-of-Experts (MoE) improve multi-task learning.
We devise task-aware gating functions to route examples from different tasks to specialized experts.
This results in a sparsely activated multi-task model with a large number of parameters, but with the same computational cost as that of a dense model.
arXiv Detail & Related papers (2022-04-16T00:56:12Z) - On the relationship between disentanglement and multi-task learning [62.997667081978825]
We take a closer look at the relationship between disentanglement and multi-task learning based on hard parameter sharing.
We show that disentanglement appears naturally during the process of multi-task neural network training.
arXiv Detail & Related papers (2021-10-07T14:35:34Z) - Rethinking Hard-Parameter Sharing in Multi-Task Learning [20.792654758645302]
Hard parameter sharing in multi-task learning (MTL) allows tasks to share some of model parameters, reducing storage cost and improving prediction accuracy.
The common sharing practice is to share bottom layers of a deep neural network among tasks while using separate top layers for each task.
Using separate bottom-layer parameters could achieve significantly better performance than the common practice.
arXiv Detail & Related papers (2021-07-23T17:26:40Z) - Exploring Relational Context for Multi-Task Dense Prediction [76.86090370115]
We consider a multi-task environment for dense prediction tasks, represented by a common backbone and independent task-specific heads.
We explore various attention-based contexts, such as global and local, in the multi-task setting.
We propose an Adaptive Task-Relational Context module, which samples the pool of all available contexts for each task pair.
arXiv Detail & Related papers (2021-04-28T16:45:56Z) - Conditional Meta-Learning of Linear Representations [57.90025697492041]
Standard meta-learning for representation learning aims to find a common representation to be shared across multiple tasks.
In this work we overcome this issue by inferring a conditioning function, mapping the tasks' side information into a representation tailored to the task at hand.
We propose a meta-algorithm capable of leveraging this advantage in practice.
arXiv Detail & Related papers (2021-03-30T12:02:14Z) - Recurrent Interaction Network for Jointly Extracting Entities and
Classifying Relations [45.79634026256055]
We design a multi-task learning model which allows the learning of interactions dynamically.
Empirical studies on two real-world datasets confirm the superiority of the proposed model.
arXiv Detail & Related papers (2020-05-01T01:03:16Z) - Multi-Task Reinforcement Learning with Soft Modularization [25.724764855681137]
Multi-task learning is a very challenging problem in reinforcement learning.
We introduce an explicit modularization technique on policy representation to alleviate this optimization issue.
We show our method improves both sample efficiency and performance over strong baselines by a large margin.
arXiv Detail & Related papers (2020-03-30T17:47:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.