How to Train Your CheXDragon: Training Chest X-Ray Models for Transfer
to Novel Tasks and Healthcare Systems
- URL: http://arxiv.org/abs/2305.08017v1
- Date: Sat, 13 May 2023 22:33:09 GMT
- Title: How to Train Your CheXDragon: Training Chest X-Ray Models for Transfer
to Novel Tasks and Healthcare Systems
- Authors: Cara Van Uden and Jeremy Irvin and Mars Huang and Nathan Dean and
Jason Carr and Andrew Ng and Curtis Langlotz
- Abstract summary: Self-supervised learning (SSL) enables label efficient training for machine learning models.
In this work, we systematically experiment with a variety of supervised and self-supervised pretraining strategies.
We show that multimodal SSL gives substantial gains over unimodal SSL in performance across new healthcare systems and tasks.
- Score: 0.118749525824656
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Self-supervised learning (SSL) enables label efficient training for machine
learning models. This is essential for domains such as medical imaging, where
labels are costly and time-consuming to curate. However, the most effective
supervised or SSL strategy for transferring models to different healthcare
systems or novel tasks is not well understood. In this work, we systematically
experiment with a variety of supervised and self-supervised pretraining
strategies using multimodal datasets of medical images (chest X-rays) and text
(radiology reports). We then evaluate their performance on data from two
external institutions with diverse sets of tasks. In addition, we experiment
with different transfer learning strategies to effectively adapt these
pretrained models to new tasks and healthcare systems. Our empirical results
suggest that multimodal SSL gives substantial gains over unimodal SSL in
performance across new healthcare systems and tasks, comparable to models
pretrained with full supervision. We demonstrate additional performance gains
with models further adapted to the new dataset and task, using multimodal
domain-adaptive pretraining (DAPT), linear probing then finetuning (LP-FT), and
both methods combined. We offer suggestions for alternative models to use in
scenarios where not all of these additions are feasible. Our results provide
guidance for improving the generalization of medical image interpretation
models to new healthcare systems and novel tasks.
Related papers
- Can LLMs' Tuning Methods Work in Medical Multimodal Domain? [14.659849302397433]
Large Language Models (LLMs) excel in world knowledge understanding, adapting them to specific subfields requires precise adjustments.
New Parameters-Efficient Fine-Tuning (PEFT) methods have emerged and achieved remarkable success in both LLMs and Large Vision-Language Models (LVLMs)
Can the fine-tuning methods for large models be transferred to the medical field to enhance transfer learning efficiency?
arXiv Detail & Related papers (2024-03-11T03:38:48Z) - Diffusion-Based Neural Network Weights Generation [80.89706112736353]
D2NWG is a diffusion-based neural network weights generation technique that efficiently produces high-performing weights for transfer learning.
Our method extends generative hyper-representation learning to recast the latent diffusion paradigm for neural network weights generation.
Our approach is scalable to large architectures such as large language models (LLMs), overcoming the limitations of current parameter generation techniques.
arXiv Detail & Related papers (2024-02-28T08:34:23Z) - LLM-Assisted Multi-Teacher Continual Learning for Visual Question Answering in Robotic Surgery [57.358568111574314]
Patient data privacy often restricts the availability of old data when updating the model.
Prior CL studies overlooked two vital problems in the surgical domain.
This paper proposes addressing these problems with a multimodal large language model (LLM) and an adaptive weight assignment methodology.
arXiv Detail & Related papers (2024-02-26T15:35:24Z) - Online Transfer Learning for RSV Case Detection [6.3076606245690385]
We introduce Multi-Source Adaptive Weighting (MSAW), an online multi-source transfer learning method.
MSAW integrates a dynamic weighting mechanism into an ensemble framework, enabling automatic adjustment of weights.
We demonstrate the effectiveness of MSAW by applying it to detect Respiratory Syncytial Virus cases within Emergency Department visits.
arXiv Detail & Related papers (2024-02-03T02:13:08Z) - MUSCLE: Multi-task Self-supervised Continual Learning to Pre-train Deep
Models for X-ray Images of Multiple Body Parts [63.30352394004674]
Multi-task Self-super-vised Continual Learning (MUSCLE) is a novel self-supervised pre-training pipeline for medical imaging tasks.
MUSCLE aggregates X-rays collected from multiple body parts for representation learning, and adopts a well-designed continual learning procedure.
We evaluate MUSCLE using 9 real-world X-ray datasets with various tasks, including pneumonia classification, skeletal abnormality classification, lung segmentation, and tuberculosis (TB) detection.
arXiv Detail & Related papers (2023-10-03T12:19:19Z) - Understanding the Tricks of Deep Learning in Medical Image Segmentation:
Challenges and Future Directions [66.40971096248946]
In this paper, we collect a series of MedISeg tricks for different model implementation phases.
We experimentally explore the effectiveness of these tricks on consistent baselines.
We also open-sourced a strong MedISeg repository, where each component has the advantage of plug-and-play.
arXiv Detail & Related papers (2022-09-21T12:30:05Z) - BERT WEAVER: Using WEight AVERaging to enable lifelong learning for
transformer-based models in biomedical semantic search engines [49.75878234192369]
We present WEAVER, a simple, yet efficient post-processing method that infuses old knowledge into the new model.
We show that applying WEAVER in a sequential manner results in similar word embedding distributions as doing a combined training on all data at once.
arXiv Detail & Related papers (2022-02-21T10:34:41Z) - A Multi-Stage Attentive Transfer Learning Framework for Improving
COVID-19 Diagnosis [49.3704402041314]
We propose a multi-stage attentive transfer learning framework for improving COVID-19 diagnosis.
Our proposed framework consists of three stages to train accurate diagnosis models through learning knowledge from multiple source tasks and data of different domains.
Importantly, we propose a novel self-supervised learning method to learn multi-scale representations for lung CT images.
arXiv Detail & Related papers (2021-01-14T01:39:19Z) - MultiMix: Sparingly Supervised, Extreme Multitask Learning From Medical
Images [13.690075845927606]
We propose a novel multitask learning model, namely MultiMix, which jointly learns disease classification and anatomical segmentation in a sparingly supervised manner.
Our experiments justify the effectiveness of our multitasking model for the classification of pneumonia and segmentation of lungs from chest X-ray images.
arXiv Detail & Related papers (2020-10-28T03:47:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.