Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image
Segmentation of Head and Neck Cancer
- URL: http://arxiv.org/abs/2305.18948v2
- Date: Wed, 2 Aug 2023 07:49:41 GMT
- Title: Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image
Segmentation of Head and Neck Cancer
- Authors: Numan Saeed, Muhammad Ridzuan, Roba Al Majzoub, Mohammad Yaqub
- Abstract summary: This paper proposes a novel fine-tuning strategy for adapting a pretrained transformer-based segmentation model on data from a new medical center.
Our strategy delivers great accuracy with minimum re-training on new-center data, significantly decreasing the computational and time costs.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Medical image segmentation is a vital healthcare endeavor requiring precise
and efficient models for appropriate diagnosis and treatment. Vision
transformer (ViT)-based segmentation models have shown great performance in
accomplishing this task. However, to build a powerful backbone, the
self-attention block of ViT requires large-scale pre-training data. The present
method of modifying pre-trained models entails updating all or some of the
backbone parameters. This paper proposes a novel fine-tuning strategy for
adapting a pretrained transformer-based segmentation model on data from a new
medical center. This method introduces a small number of learnable parameters,
termed prompts, into the input space (less than 1\% of model parameters) while
keeping the rest of the model parameters frozen. Extensive studies employing
data from new unseen medical centers show that the prompt-based fine-tuning of
medical segmentation models provides excellent performance regarding the
new-center data with a negligible drop regarding the old centers. Additionally,
our strategy delivers great accuracy with minimum re-training on new-center
data, significantly decreasing the computational and time costs of fine-tuning
pre-trained models.
Related papers
- Low-Rank Continual Pyramid Vision Transformer: Incrementally Segment Whole-Body Organs in CT with Light-Weighted Adaptation [10.746776960260297]
We propose a new continual whole-body organ segmentation model with light-weighted low-rank adaptation (LoRA)
We first train and freeze a pyramid vision transformer (PVT) base segmentation model on the initial task, then continually add light-weighted trainable LoRA parameters to the frozen model for each new learning task.
Our proposed model continually segments new organs without catastrophic forgetting and meanwhile maintaining a low parameter increasing rate.
arXiv Detail & Related papers (2024-10-07T02:00:13Z) - SaRA: High-Efficient Diffusion Model Fine-tuning with Progressive Sparse Low-Rank Adaptation [52.6922833948127]
In this work, we investigate the importance of parameters in pre-trained diffusion models.
We propose a novel model fine-tuning method to make full use of these ineffective parameters.
Our method enhances the generative capabilities of pre-trained models in downstream applications.
arXiv Detail & Related papers (2024-09-10T16:44:47Z) - Few-Shot Airway-Tree Modeling using Data-Driven Sparse Priors [0.0]
Few-shot learning approaches are cost-effective to transfer pre-trained models using only limited annotated data.
We train a data-driven sparsification module to enhance airways efficiently in lung CT scans.
We then incorporate these sparse representations in a standard supervised segmentation pipeline as a pretraining step to enhance the performance of the DL models.
arXiv Detail & Related papers (2024-07-05T13:46:11Z) - DVPT: Dynamic Visual Prompt Tuning of Large Pre-trained Models for
Medical Image Analysis [30.608225734194416]
We propose a dynamic visual prompt tuning method, named DVPT, for medical image analysis.
It can extract knowledge beneficial to downstream tasks from large models with a few trainable parameters.
It can save up to 60% labeled data and 99% storage cost of ViT-B/16.
arXiv Detail & Related papers (2023-07-19T07:11:11Z) - Learnable Weight Initialization for Volumetric Medical Image Segmentation [66.3030435676252]
We propose a learnable weight-based hybrid medical image segmentation approach.
Our approach is easy to integrate into any hybrid model and requires no external training data.
Experiments on multi-organ and lung cancer segmentation tasks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-06-15T17:55:05Z) - Towards Foundation Models and Few-Shot Parameter-Efficient Fine-Tuning for Volumetric Organ Segmentation [20.94974284175104]
Few-Shot Efficient Fine-Tuning (FSEFT) is a novel and realistic scenario for adapting medical image segmentation foundation models.
Our comprehensive transfer learning experiments confirm the suitability of foundation models in medical image segmentation and unveil the limitations of popular fine-tuning strategies in few-shot scenarios.
arXiv Detail & Related papers (2023-03-29T22:50:05Z) - Prompt Tuning for Parameter-efficient Medical Image Segmentation [79.09285179181225]
We propose and investigate several contributions to achieve a parameter-efficient but effective adaptation for semantic segmentation on two medical imaging datasets.
We pre-train this architecture with a dedicated dense self-supervision scheme based on assignments to online generated prototypes.
We demonstrate that the resulting neural network model is able to attenuate the gap between fully fine-tuned and parameter-efficiently adapted models.
arXiv Detail & Related papers (2022-11-16T21:55:05Z) - Self-Distillation for Further Pre-training of Transformers [83.84227016847096]
We propose self-distillation as a regularization for a further pre-training stage.
We empirically validate the efficacy of self-distillation on a variety of benchmark datasets for image and text classification tasks.
arXiv Detail & Related papers (2022-09-30T02:25:12Z) - Kronecker Factorization for Preventing Catastrophic Forgetting in
Large-scale Medical Entity Linking [7.723047334864811]
In the medical domain, sequential training on tasks may sometimes be the only way to train models.
catastrophic forgetting, i.e., a substantial drop in accuracy on prior tasks when a model is updated for a new task.
We show the effectiveness of this technique on the important and illustrative task of medical entity linking across three datasets.
arXiv Detail & Related papers (2021-11-11T01:51:01Z) - Self-Training with Improved Regularization for Sample-Efficient Chest
X-Ray Classification [80.00316465793702]
We present a deep learning framework that enables robust modeling in challenging scenarios.
Our results show that using 85% lesser labeled data, we can build predictive models that match the performance of classifiers trained in a large-scale data setting.
arXiv Detail & Related papers (2020-05-03T02:36:00Z) - Automatic Data Augmentation via Deep Reinforcement Learning for
Effective Kidney Tumor Segmentation [57.78765460295249]
We develop a novel automatic learning-based data augmentation method for medical image segmentation.
In our method, we innovatively combine the data augmentation module and the subsequent segmentation module in an end-to-end training manner with a consistent loss.
We extensively evaluated our method on CT kidney tumor segmentation which validated the promising results of our method.
arXiv Detail & Related papers (2020-02-22T14:10:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.