MTI-Net: Multi-Scale Task Interaction Networks for Multi-Task Learning
- URL: http://arxiv.org/abs/2001.06902v5
- Date: Wed, 8 Jul 2020 19:58:22 GMT
- Title: MTI-Net: Multi-Scale Task Interaction Networks for Multi-Task Learning
- Authors: Simon Vandenhende, Stamatios Georgoulis and Luc Van Gool
- Abstract summary: We show that tasks with high affinity at a certain scale are not guaranteed to retain this behaviour at other scales.
We propose a novel architecture, namely MTI-Net, that builds upon this finding.
- Score: 82.62433731378455
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we argue about the importance of considering task interactions
at multiple scales when distilling task information in a multi-task learning
setup. In contrast to common belief, we show that tasks with high affinity at a
certain scale are not guaranteed to retain this behaviour at other scales, and
vice versa. We propose a novel architecture, namely MTI-Net, that builds upon
this finding in three ways. First, it explicitly models task interactions at
every scale via a multi-scale multi-modal distillation unit. Second, it
propagates distilled task information from lower to higher scales via a feature
propagation module. Third, it aggregates the refined task features from all
scales via a feature aggregation unit to produce the final per-task
predictions.
Extensive experiments on two multi-task dense labeling datasets show that,
unlike prior work, our multi-task model delivers on the full potential of
multi-task learning, that is, smaller memory footprint, reduced number of
calculations, and better performance w.r.t. single-task learning. The code is
made publicly available:
https://github.com/SimonVandenhende/Multi-Task-Learning-PyTorch.
Related papers
- Cross-Task Affinity Learning for Multitask Dense Scene Predictions [5.939164722752263]
Multitask learning (MTL) has become prominent for its ability to predict multiple tasks jointly.
We introduce the Cross-Task Affinity Learning (CTAL) module, a lightweight framework that enhances task refinement in multitask networks.
Our results demonstrate state-of-the-art MTL performance for both CNN and transformer backbones, using significantly fewer parameters than single-task learning.
arXiv Detail & Related papers (2024-01-20T05:31:47Z) - Distribution Matching for Multi-Task Learning of Classification Tasks: a
Large-Scale Study on Faces & Beyond [62.406687088097605]
Multi-Task Learning (MTL) is a framework, where multiple related tasks are learned jointly and benefit from a shared representation space.
We show that MTL can be successful with classification tasks with little, or non-overlapping annotations.
We propose a novel approach, where knowledge exchange is enabled between the tasks via distribution matching.
arXiv Detail & Related papers (2024-01-02T14:18:11Z) - YOLOR-Based Multi-Task Learning [12.5920336941241]
Multi-task learning (MTL) aims to learn multiple tasks using a single model and jointly improve all of them assuming generalization and shared semantics.
We propose building on You Only Learn One Representation (YOLOR), a network architecture specifically designed for multitasking.
We find that our method achieves competitive performance on all tasks while maintaining a low parameter count and without any pre-training.
arXiv Detail & Related papers (2023-09-29T01:42:21Z) - Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners [67.5865966762559]
We study whether sparsely activated Mixture-of-Experts (MoE) improve multi-task learning.
We devise task-aware gating functions to route examples from different tasks to specialized experts.
This results in a sparsely activated multi-task model with a large number of parameters, but with the same computational cost as that of a dense model.
arXiv Detail & Related papers (2022-04-16T00:56:12Z) - Medusa: Universal Feature Learning via Attentional Multitasking [65.94499390875046]
Recent approaches to multi-task learning have focused on modelling connections between tasks at the decoder level.
We argue that MTL is a stepping stone towards universal feature learning (UFL), which is the ability to learn generic features that can be applied to new tasks without retraining.
We show the effectiveness of Medusa in UFL (+13.18% improvement) while maintaining MTL performance and being 25% more efficient than previous approaches.
arXiv Detail & Related papers (2022-04-12T10:52:28Z) - Attentive Task Interaction Network for Multi-Task Learning [4.1372815372396525]
ATI-Net employs knowledge distillation of the latent features for each task, then combines the feature maps to provide improved contextualized information to the decoder.
This novel approach to introducing knowledge distillation into an attention based multitask network outperforms state of the art MTL baselines.
arXiv Detail & Related papers (2022-01-25T22:03:20Z) - Reparameterizing Convolutions for Incremental Multi-Task Learning
without Task Interference [75.95287293847697]
Two common challenges in developing multi-task models are often overlooked in literature.
First, enabling the model to be inherently incremental, continuously incorporating information from new tasks without forgetting the previously learned ones (incremental learning)
Second, eliminating adverse interactions amongst tasks, which has been shown to significantly degrade the single-task performance in a multi-task setup (task interference)
arXiv Detail & Related papers (2020-07-24T14:44:46Z) - Knowledge Distillation for Multi-task Learning [38.20005345733544]
Multi-task learning (MTL) is to learn one single model that performs multiple tasks for achieving good performance on all tasks and lower cost on computation.
Learning such a model requires to jointly optimize losses of a set of tasks with different difficulty levels, magnitudes, and characteristics.
We propose a knowledge distillation based method in this work to address the imbalance problem in multi-task learning.
arXiv Detail & Related papers (2020-07-14T08:02:42Z) - Multi-Task Learning for Dense Prediction Tasks: A Survey [87.66280582034838]
Multi-task learning (MTL) techniques have shown promising results w.r.t. performance, computations and/or memory footprint.
We provide a well-rounded view on state-of-the-art deep learning approaches for MTL in computer vision.
arXiv Detail & Related papers (2020-04-28T09:15:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.