Emerging Relation Network and Task Embedding for Multi-Task Regression
Problems
- URL: http://arxiv.org/abs/2004.14034v1
- Date: Wed, 29 Apr 2020 09:02:24 GMT
- Title: Emerging Relation Network and Task Embedding for Multi-Task Regression
Problems
- Authors: Jens Schreiber, Bernhard Sick
- Abstract summary: Multi-task learning (mtl) provides state-of-the-art results in many applications of computer vision and natural language processing.
This article provides a comparative study of the following recent and important mtl architectures.
We introduce a new mtl architecture named emerging relation network (ern) which can be considered as an extension of the sluice network.
- Score: 5.953831950062808
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-task learning (mtl) provides state-of-the-art results in many
applications of computer vision and natural language processing. In contrast to
single-task learning (stl), mtl allows for leveraging knowledge between related
tasks improving prediction results on the main task (in contrast to an
auxiliary task) or all tasks. However, there is a limited number of comparative
studies on applying mtl architectures for regression and time series problems
taking recent advances of mtl into account. An interesting, non-linear problem
is the forecast of the expected power generation for renewable power plants.
Therefore, this article provides a comparative study of the following recent
and important mtl architectures: Hard parameter sharing, cross-stitch network,
sluice network (sn). They are compared to a multi-layer perceptron model of
similar size in an stl setting. Additionally, we provide a simple, yet
effective approach to model task specific information through an embedding
layer in an multi-layer perceptron, referred to as task embedding. Further, we
introduce a new mtl architecture named emerging relation network (ern), which
can be considered as an extension of the sluice network. For a solar power
dataset, the task embedding achieves the best mean improvement with 14.9%. The
mean improvement of the ern and the sn on the solar dataset is of similar
magnitude with 14.7% and 14.8%. On a wind power dataset, only the ern achieves
a significant improvement of up to 7.7%. Results suggest that the ern is
beneficial when tasks are only loosely related and the prediction problem is
more non-linear. Contrary, the proposed task embedding is advantageous when
tasks are strongly correlated. Further, the task embedding provides an
effective approach with reduced computational effort compared to other mtl
architectures.
Related papers
- Cross-Task Affinity Learning for Multitask Dense Scene Predictions [5.939164722752263]
Multitask learning (MTL) has become prominent for its ability to predict multiple tasks jointly.
We introduce the Cross-Task Affinity Learning (CTAL) module, a lightweight framework that enhances task refinement in multitask networks.
Our results demonstrate state-of-the-art MTL performance for both CNN and transformer backbones, using significantly fewer parameters than single-task learning.
arXiv Detail & Related papers (2024-01-20T05:31:47Z) - Distribution Matching for Multi-Task Learning of Classification Tasks: a
Large-Scale Study on Faces & Beyond [62.406687088097605]
Multi-Task Learning (MTL) is a framework, where multiple related tasks are learned jointly and benefit from a shared representation space.
We show that MTL can be successful with classification tasks with little, or non-overlapping annotations.
We propose a novel approach, where knowledge exchange is enabled between the tasks via distribution matching.
arXiv Detail & Related papers (2024-01-02T14:18:11Z) - Multi-task neural networks by learned contextual inputs [0.0]
It is a multi-task learning architecture based on a fully shared neural network and an augmented input vector containing trainable task parameters.
The architecture is interesting due to its powerful task mechanism, which facilitates a low-dimensional task parameter space.
The architecture's performance is compared to similar neural network architectures on ten datasets.
arXiv Detail & Related papers (2023-03-01T19:25:52Z) - Arch-Graph: Acyclic Architecture Relation Predictor for
Task-Transferable Neural Architecture Search [96.31315520244605]
Arch-Graph is a transferable NAS method that predicts task-specific optimal architectures.
We show Arch-Graph's transferability and high sample efficiency across numerous tasks.
It is able to find top 0.16% and 0.29% architectures on average on two search spaces under the budget of only 50 models.
arXiv Detail & Related papers (2022-04-12T16:46:06Z) - Task Adaptive Parameter Sharing for Multi-Task Learning [114.80350786535952]
Adaptive Task Adapting Sharing (TAPS) is a method for tuning a base model to a new task by adaptively modifying a small, task-specific subset of layers.
Compared to other methods, TAPS retains high accuracy on downstream tasks while introducing few task-specific parameters.
We evaluate our method on a suite of fine-tuning tasks and architectures (ResNet, DenseNet, ViT) and show that it achieves state-of-the-art performance while being simple to implement.
arXiv Detail & Related papers (2022-03-30T23:16:07Z) - On-edge Multi-task Transfer Learning: Model and Practice with
Data-driven Task Allocation [20.20889051697198]
We show that task allocation with task importance for Multi-task Transfer Learning (MTL) is a variant of the NP-complete Knapsack problem.
We propose a Data-driven Cooperative Task Allocation (DCTA) approach to solve TATIM with high computational efficiency.
Our DCTA reduces 3.24 times of processing time, and saves 48.4% energy consumption compared with the state-of-the-art when solving TATIM.
arXiv Detail & Related papers (2021-07-06T08:24:25Z) - Multi-task Over-the-Air Federated Learning: A Non-Orthogonal
Transmission Approach [52.85647632037537]
We propose a multi-task over-theair federated learning (MOAFL) framework, where multiple learning tasks share edge devices for data collection and learning models under the coordination of a edge server (ES)
Both the convergence analysis and numerical results demonstrate that the MOAFL framework can significantly reduce the uplink bandwidth consumption of multiple tasks without causing substantial learning performance degradation.
arXiv Detail & Related papers (2021-06-27T13:09:32Z) - Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning
in NLP Using Fewer Parameters & Less Data [5.689320790746046]
Multi-Task Learning (MTL) networks have emerged as a promising method for transferring learned knowledge across different tasks.
However, MTL must deal with challenges such as: overfitting to low resource tasks, catastrophic forgetting, and negative task transfer.
We propose a novel Transformer architecture consisting of a new conditional attention mechanism and a set of task-conditioned modules.
arXiv Detail & Related papers (2020-09-19T02:04:34Z) - Reparameterizing Convolutions for Incremental Multi-Task Learning
without Task Interference [75.95287293847697]
Two common challenges in developing multi-task models are often overlooked in literature.
First, enabling the model to be inherently incremental, continuously incorporating information from new tasks without forgetting the previously learned ones (incremental learning)
Second, eliminating adverse interactions amongst tasks, which has been shown to significantly degrade the single-task performance in a multi-task setup (task interference)
arXiv Detail & Related papers (2020-07-24T14:44:46Z) - Task-Feature Collaborative Learning with Application to Personalized
Attribute Prediction [166.87111665908333]
We propose a novel multi-task learning method called Task-Feature Collaborative Learning (TFCL)
Specifically, we first propose a base model with a heterogeneous block-diagonal structure regularizer to leverage the collaborative grouping of features and tasks.
As a practical extension, we extend the base model by allowing overlapping features and differentiating the hard tasks.
arXiv Detail & Related papers (2020-04-29T02:32:04Z) - Deeper Task-Specificity Improves Joint Entity and Relation Extraction [0.0]
Multi-task learning (MTL) is an effective method for learning related tasks, but designing MTL models requires deciding which and how many parameters should be task-specific.
We propose a novel neural architecture that allows for deeper task-specificity than does prior work.
arXiv Detail & Related papers (2020-02-15T18:34:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.