A Comprehensive Evaluation of Multi-task Learning and Multi-task
Pre-training on EHR Time-series Data
- URL: http://arxiv.org/abs/2007.10185v1
- Date: Mon, 20 Jul 2020 15:19:28 GMT
- Title: A Comprehensive Evaluation of Multi-task Learning and Multi-task
Pre-training on EHR Time-series Data
- Authors: Matthew B.A. McDermott (1), Bret Nestor (2), Evan Kim (1), Wancong
Zhang (3), Anna Goldenberg (2, 4, 5), Peter Szolovits (1), Marzyeh Ghassemi
(2, 4) ((1) CSAIL, MIT, (2) University of Toronto, (3) NYU, (4) Vector
Institute, (5) SickKids)
- Abstract summary: Multi-task learning (MTL) is a machine learning technique aiming to improve model performance by leveraging information across many tasks.
In this work, we examine MTL across a battery of tasks on EHR time-series data.
We find that while MTL does suffer from common negative transfer, we can realize significant gains via MTL pre-training combined with single-task fine-tuning.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-task learning (MTL) is a machine learning technique aiming to improve
model performance by leveraging information across many tasks. It has been used
extensively on various data modalities, including electronic health record
(EHR) data. However, despite significant use on EHR data, there has been little
systematic investigation of the utility of MTL across the diverse set of
possible tasks and training schemes of interest in healthcare. In this work, we
examine MTL across a battery of tasks on EHR time-series data. We find that
while MTL does suffer from common negative transfer, we can realize significant
gains via MTL pre-training combined with single-task fine-tuning. We
demonstrate that these gains can be achieved in a task-independent manner and
offer not only minor improvements under traditional learning, but also notable
gains in a few-shot learning context, thereby suggesting this could be a
scalable vehicle to offer improved performance in important healthcare
contexts.
Related papers
- Data-CUBE: Data Curriculum for Instruction-based Sentence Representation
Learning [85.66907881270785]
We propose a data curriculum method, namely Data-CUBE, that arranges the orders of all the multi-task data for training.
In the task level, we aim to find the optimal task order to minimize the total cross-task interference risk.
In the instance level, we measure the difficulty of all instances per task, then divide them into the easy-to-difficult mini-batches for training.
arXiv Detail & Related papers (2024-01-07T18:12:20Z) - Distribution Matching for Multi-Task Learning of Classification Tasks: a
Large-Scale Study on Faces & Beyond [62.406687088097605]
Multi-Task Learning (MTL) is a framework, where multiple related tasks are learned jointly and benefit from a shared representation space.
We show that MTL can be successful with classification tasks with little, or non-overlapping annotations.
We propose a novel approach, where knowledge exchange is enabled between the tasks via distribution matching.
arXiv Detail & Related papers (2024-01-02T14:18:11Z) - Equitable Multi-task Learning [18.65048321820911]
Multi-task learning (MTL) has achieved great success in various research domains, such as CV, NLP and IR.
We propose a novel multi-task optimization method, named EMTL, to achieve equitable MTL.
Our method stably outperforms state-of-the-art methods on the public benchmark datasets of two different research domains.
arXiv Detail & Related papers (2023-06-15T03:37:23Z) - Diffusion Model is an Effective Planner and Data Synthesizer for
Multi-Task Reinforcement Learning [101.66860222415512]
Multi-Task Diffusion Model (textscMTDiff) is a diffusion-based method that incorporates Transformer backbones and prompt learning for generative planning and data synthesis.
For generative planning, we find textscMTDiff outperforms state-of-the-art algorithms across 50 tasks on Meta-World and 8 maps on Maze2D.
arXiv Detail & Related papers (2023-05-29T05:20:38Z) - When Does Aggregating Multiple Skills with Multi-Task Learning Work? A
Case Study in Financial NLP [22.6364117325639]
Multi-task learning (MTL) aims at achieving a better model by leveraging data and knowledge from multiple tasks.
Our findings suggest that the key to MTL success lies in skill diversity, relatedness between tasks, and choice of aggregation size and shared capacity.
arXiv Detail & Related papers (2023-05-23T12:37:14Z) - Multi-Task Instruction Tuning of LLaMa for Specific Scenarios: A
Preliminary Study on Writing Assistance [60.40541387785977]
Small foundational models can display remarkable proficiency in tackling diverse tasks when fine-tuned using instruction-driven data.
In this work, we investigate a practical problem setting where the primary focus is on one or a few particular tasks rather than general-purpose instruction following.
Experimental results show that fine-tuning LLaMA on writing instruction data significantly improves its ability on writing tasks.
arXiv Detail & Related papers (2023-05-22T16:56:44Z) - Task Compass: Scaling Multi-task Pre-training with Task Prefix [122.49242976184617]
Existing studies show that multi-task learning with large-scale supervised tasks suffers from negative effects across tasks.
We propose a task prefix guided multi-task pre-training framework to explore the relationships among tasks.
Our model can not only serve as the strong foundation backbone for a wide range of tasks but also be feasible as a probing tool for analyzing task relationships.
arXiv Detail & Related papers (2022-10-12T15:02:04Z) - When to Use Multi-Task Learning vs Intermediate Fine-Tuning for
Pre-Trained Encoder Transfer Learning [15.39115079099451]
Transfer learning (TL) in natural language processing has seen a surge of interest in recent years.
Three main strategies have emerged for making use of multiple supervised datasets during fine-tuning.
We compare all three TL methods in a comprehensive analysis on the GLUE dataset suite.
arXiv Detail & Related papers (2022-05-17T06:48:45Z) - Task Uncertainty Loss Reduce Negative Transfer in Asymmetric Multi-task
Feature Learning [0.0]
Multi-task learning (MTL) can improve task performance overall relative to single-task learning (STL), but can hide negative transfer (NT)
Asymmetric multitask feature learning (AMTFL) is an approach that tries to address this by allowing tasks with higher loss values to have smaller influence on feature representations for learning other tasks.
We present examples of NT in two datasets (image recognition and pharmacogenomics) and tackle this challenge by using aleatoric homoscedastic uncertainty to capture the relative confidence between tasks, and set weights for task loss.
arXiv Detail & Related papers (2020-12-17T13:30:45Z) - A Brief Review of Deep Multi-task Learning and Auxiliary Task Learning [0.0]
Multi-task learning (MTL) optimize several learning tasks simultaneously.
Auxiliary tasks can be added to the main task to boost the performance.
arXiv Detail & Related papers (2020-07-02T14:23:39Z) - Learning to Multi-Task Learn for Better Neural Machine Translation [53.06405021125476]
Multi-task learning is an elegant approach to inject linguistic-related biases into neural machine translation models.
We propose a novel framework for learning the training schedule, ie learning to multi-task learn, for the biased-MTL setting of interest.
Experiments show the resulting automatically learned training schedulers are competitive with the best, and lead to up to +1.1 BLEU score improvements.
arXiv Detail & Related papers (2020-01-10T03:12:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.