Temporally Correlated Task Scheduling for Sequence Learning
- URL: http://arxiv.org/abs/2007.05290v2
- Date: Fri, 2 Jul 2021 12:39:00 GMT
- Title: Temporally Correlated Task Scheduling for Sequence Learning
- Authors: Xueqing Wu, Lewen Wang, Yingce Xia, Weiqing Liu, Lijun Wu, Shufang
Xie, Tao Qin, Tie-Yan Liu
- Abstract summary: In many applications, a sequence learning task is usually associated with multiple temporally correlated auxiliary tasks.
We introduce a learnable scheduler to sequence learning, which can adaptively select auxiliary tasks for training.
Our method significantly improves the performance of simultaneous machine translation and stock trend forecasting.
- Score: 143.70523777803723
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sequence learning has attracted much research attention from the machine
learning community in recent years. In many applications, a sequence learning
task is usually associated with multiple temporally correlated auxiliary tasks,
which are different in terms of how much input information to use or which
future step to predict. For example, (i) in simultaneous machine translation,
one can conduct translation under different latency (i.e., how many input words
to read/wait before translation); (ii) in stock trend forecasting, one can
predict the price of a stock in different future days (e.g., tomorrow, the day
after tomorrow). While it is clear that those temporally correlated tasks can
help each other, there is a very limited exploration on how to better leverage
multiple auxiliary tasks to boost the performance of the main task. In this
work, we introduce a learnable scheduler to sequence learning, which can
adaptively select auxiliary tasks for training depending on the model status
and the current training data. The scheduler and the model for the main task
are jointly trained through bi-level optimization. Experiments show that our
method significantly improves the performance of simultaneous machine
translation and stock trend forecasting.
Related papers
- Pre-training Multi-task Contrastive Learning Models for Scientific
Literature Understanding [52.723297744257536]
Pre-trained language models (LMs) have shown effectiveness in scientific literature understanding tasks.
We propose a multi-task contrastive learning framework, SciMult, to facilitate common knowledge sharing across different literature understanding tasks.
arXiv Detail & Related papers (2023-05-23T16:47:22Z) - Learning Easily Updated General Purpose Text Representations with
Adaptable Task-Specific Prefixes [22.661527526471996]
Fine-tuning a large pre-trained language model for each downstream task causes computational burdens.
We propose a prefix-based method to learn the fixed text representations with source tasks.
arXiv Detail & Related papers (2023-05-22T21:31:03Z) - Effective Cross-Task Transfer Learning for Explainable Natural Language
Inference with T5 [50.574918785575655]
We compare sequential fine-tuning with a model for multi-task learning in the context of boosting performance on two tasks.
Our results show that while sequential multi-task learning can be tuned to be good at the first of two target tasks, it performs less well on the second and additionally struggles with overfitting.
arXiv Detail & Related papers (2022-10-31T13:26:08Z) - Task Compass: Scaling Multi-task Pre-training with Task Prefix [122.49242976184617]
Existing studies show that multi-task learning with large-scale supervised tasks suffers from negative effects across tasks.
We propose a task prefix guided multi-task pre-training framework to explore the relationships among tasks.
Our model can not only serve as the strong foundation backbone for a wide range of tasks but also be feasible as a probing tool for analyzing task relationships.
arXiv Detail & Related papers (2022-10-12T15:02:04Z) - On Steering Multi-Annotations per Sample for Multi-Task Learning [79.98259057711044]
The study of multi-task learning has drawn great attention from the community.
Despite the remarkable progress, the challenge of optimally learning different tasks simultaneously remains to be explored.
Previous works attempt to modify the gradients from different tasks. Yet these methods give a subjective assumption of the relationship between tasks, and the modified gradient may be less accurate.
In this paper, we introduce Task Allocation(STA), a mechanism that addresses this issue by a task allocation approach, in which each sample is randomly allocated a subset of tasks.
For further progress, we propose Interleaved Task Allocation(ISTA) to iteratively allocate all
arXiv Detail & Related papers (2022-03-06T11:57:18Z) - Transfer Learning in Conversational Analysis through Reusing
Preprocessing Data as Supervisors [52.37504333689262]
Using noisy labels in single-task learning increases the risk of over-fitting.
Auxiliary tasks could improve the performance of the primary task learning during the same training.
arXiv Detail & Related papers (2021-12-02T08:40:42Z) - Learning Multiple Dense Prediction Tasks from Partially Annotated Data [41.821234589075445]
We look at jointly learning of multiple dense prediction tasks on partially annotated data, which we call multi-task partially-supervised learning.
We propose a multi-task training procedure that successfully leverages task relations to supervise its multi-task learning when data is partially annotated.
We rigorously demonstrate that our proposed method effectively exploits the images with unlabelled tasks and outperforms existing semi-supervised learning approaches and related methods on three standard benchmarks.
arXiv Detail & Related papers (2021-11-29T19:03:12Z) - Multi-Task Time Series Forecasting With Shared Attention [15.294939035413217]
We propose two self-attention based sharing schemes for multi-task time series forecasting.
Our proposed architectures can not only outperform the state-of-the-art single-task forecasting baselines but also outperform the RNN-based multi-task forecasting method.
arXiv Detail & Related papers (2021-01-24T04:25:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.