MedSegDiff-V2: Diffusion based Medical Image Segmentation with
Transformer
- URL: http://arxiv.org/abs/2301.11798v2
- Date: Sun, 24 Dec 2023 03:02:12 GMT
- Title: MedSegDiff-V2: Diffusion based Medical Image Segmentation with
Transformer
- Authors: Junde Wu, Wei Ji, Huazhu Fu, Min Xu, Yueming Jin, Yanwu Xu
- Abstract summary: We propose a novel Transformer-based Diffusion framework, called MedSegDiff-V2.
We verify its effectiveness on 20 medical image segmentation tasks with different image modalities.
- Score: 53.575573940055335
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The Diffusion Probabilistic Model (DPM) has recently gained popularity in the
field of computer vision, thanks to its image generation applications, such as
Imagen, Latent Diffusion Models, and Stable Diffusion, which have demonstrated
impressive capabilities and sparked much discussion within the community.
Recent investigations have further unveiled the utility of DPM in the domain of
medical image analysis, as underscored by the commendable performance exhibited
by the medical image segmentation model across various tasks. Although these
models were originally underpinned by a UNet architecture, there exists a
potential avenue for enhancing their performance through the integration of
vision transformer mechanisms. However, we discovered that simply combining
these two models resulted in subpar performance. To effectively integrate these
two cutting-edge techniques for the Medical image segmentation, we propose a
novel Transformer-based Diffusion framework, called MedSegDiff-V2. We verify
its effectiveness on 20 medical image segmentation tasks with different image
modalities. Through comprehensive evaluation, our approach demonstrates
superiority over prior state-of-the-art (SOTA) methodologies. Code is released
at https://github.com/KidsWithTokens/MedSegDiff
Related papers
- MedVisionLlama: Leveraging Pre-Trained Large Language Model Layers to Enhance Medical Image Segmentation [0.8437187555622164]
This study explores enhancing Vision Transformers (ViTs) for medical image segmentation by integrating pre-trained LLM transformer blocks.
Our approach, which incorporates a frozen LLM transformer block into the encoder of a ViT-based model, leads to substantial improvements in segmentation performance.
The enhanced model shows significant performance gains, including an average Dice score increase from 0.74 to 0.79 and improvements in accuracy, precision, and the Jaccard Index.
arXiv Detail & Related papers (2024-10-03T14:50:33Z) - Adapting Visual-Language Models for Generalizable Anomaly Detection in Medical Images [68.42215385041114]
This paper introduces a novel lightweight multi-level adaptation and comparison framework to repurpose the CLIP model for medical anomaly detection.
Our approach integrates multiple residual adapters into the pre-trained visual encoder, enabling a stepwise enhancement of visual features across different levels.
Our experiments on medical anomaly detection benchmarks demonstrate that our method significantly surpasses current state-of-the-art models.
arXiv Detail & Related papers (2024-03-19T09:28:19Z) - Dual-scale Enhanced and Cross-generative Consistency Learning for Semi-supervised Medical Image Segmentation [49.57907601086494]
Medical image segmentation plays a crucial role in computer-aided diagnosis.
We propose a novel Dual-scale Enhanced and Cross-generative consistency learning framework for semi-supervised medical image (DEC-Seg)
arXiv Detail & Related papers (2023-12-26T12:56:31Z) - EAFP-Med: An Efficient Adaptive Feature Processing Module Based on
Prompts for Medical Image Detection [27.783012550610387]
Cross-domain adaptive medical image detection is challenging due to the differences in lesion representations across various medical imaging technologies.
We propose EAFP-Med, an efficient adaptive feature processing module based on prompts for medical image detection.
EAFP-Med can efficiently extract lesion features from various medical images based on prompts, enhancing the model's performance.
arXiv Detail & Related papers (2023-11-27T05:10:15Z) - Customizing General-Purpose Foundation Models for Medical Report
Generation [64.31265734687182]
The scarcity of labelled medical image-report pairs presents great challenges in the development of deep and large-scale neural networks.
We propose customizing off-the-shelf general-purpose large-scale pre-trained models, i.e., foundation models (FMs) in computer vision and natural language processing.
arXiv Detail & Related papers (2023-06-09T03:02:36Z) - Ambiguous Medical Image Segmentation using Diffusion Models [60.378180265885945]
We introduce a single diffusion model-based approach that produces multiple plausible outputs by learning a distribution over group insights.
Our proposed model generates a distribution of segmentation masks by leveraging the inherent sampling process of diffusion.
Comprehensive results show that our proposed approach outperforms existing state-of-the-art ambiguous segmentation networks.
arXiv Detail & Related papers (2023-04-10T17:58:22Z) - DiffMIC: Dual-Guidance Diffusion Network for Medical Image
Classification [32.67098520984195]
We propose the first diffusion-based model (named DiffMIC) to address general medical image classification.
Our experimental results demonstrate that DiffMIC outperforms state-of-the-art methods by a significant margin.
arXiv Detail & Related papers (2023-03-19T09:15:45Z) - Diff-UNet: A Diffusion Embedded Network for Volumetric Segmentation [41.608617301275935]
We propose a novel end-to-end framework, called Diff-UNet, for medical volumetric segmentation.
Our approach integrates the diffusion model into a standard U-shaped architecture to extract semantic information from the input volume effectively.
We evaluate our method on three datasets, including multimodal brain tumors in MRI, liver tumors, and multi-organ CT volumes.
arXiv Detail & Related papers (2023-03-18T04:06:18Z) - MedSegDiff: Medical Image Segmentation with Diffusion Probabilistic
Model [8.910108260704964]
Diffusion model (DPM) recently becomes one of the hottest topic in computer vision.
We propose the first DPM based model toward general medical image segmentation tasks, which we named MedSegDiff.
experimental results show that MedSegDiff outperforms state-of-the-art (SOTA) methods with considerable performance gap.
arXiv Detail & Related papers (2022-11-01T17:24:44Z) - Medical Transformer: Gated Axial-Attention for Medical Image
Segmentation [73.98974074534497]
We study the feasibility of using Transformer-based network architectures for medical image segmentation tasks.
We propose a Gated Axial-Attention model which extends the existing architectures by introducing an additional control mechanism in the self-attention module.
To train the model effectively on medical images, we propose a Local-Global training strategy (LoGo) which further improves the performance.
arXiv Detail & Related papers (2021-02-21T18:35:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.