Partition-A-Medical-Image: Extracting Multiple Representative
Sub-regions for Few-shot Medical Image Segmentation
- URL: http://arxiv.org/abs/2309.11172v1
- Date: Wed, 20 Sep 2023 09:31:57 GMT
- Title: Partition-A-Medical-Image: Extracting Multiple Representative
Sub-regions for Few-shot Medical Image Segmentation
- Authors: Yazhou Zhu, Shidong Wang, Tong Xin, Zheng Zhang, Haofeng Zhang
- Abstract summary: Few-shot Medical Image (FSMIS) is a more promising solution for medical image segmentation tasks.
We present an approach to extract multiple representative sub-regions from a given support medical image.
We then introduce a novel Prototypical Representation Debiasing (PRD) module based on a two-way elimination mechanism.
- Score: 23.926487942901872
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Few-shot Medical Image Segmentation (FSMIS) is a more promising solution for
medical image segmentation tasks where high-quality annotations are naturally
scarce. However, current mainstream methods primarily focus on extracting
holistic representations from support images with large intra-class variations
in appearance and background, and encounter difficulties in adapting to query
images. In this work, we present an approach to extract multiple representative
sub-regions from a given support medical image, enabling fine-grained selection
over the generated image regions. Specifically, the foreground of the support
image is decomposed into distinct regions, which are subsequently used to
derive region-level representations via a designed Regional Prototypical
Learning (RPL) module. We then introduce a novel Prototypical Representation
Debiasing (PRD) module based on a two-way elimination mechanism which
suppresses the disturbance of regional representations by a self-support,
Multi-direction Self-debiasing (MS) block, and a support-query, Interactive
Debiasing (ID) block. Finally, an Assembled Prediction (AP) module is devised
to balance and integrate predictions of multiple prototypical representations
learned using stacked PRD modules. Results obtained through extensive
experiments on three publicly accessible medical imaging datasets demonstrate
consistent improvements over the leading FSMIS methods. The source code is
available at https://github.com/YazhouZhu19/PAMI.
Related papers
- Unlocking the Power of Spatial and Temporal Information in Medical Multimodal Pre-training [99.2891802841936]
We introduce the Med-ST framework for fine-grained spatial and temporal modeling.
For spatial modeling, Med-ST employs the Mixture of View Expert (MoVE) architecture to integrate different visual features from both frontal and lateral views.
For temporal modeling, we propose a novel cross-modal bidirectional cycle consistency objective by forward mapping classification (FMC) and reverse mapping regression (RMR)
arXiv Detail & Related papers (2024-05-30T03:15:09Z) - FeaInfNet: Diagnosis in Medical Image with Feature-Driven Inference and
Visual Explanations [4.022446255159328]
Interpretable deep learning models have received widespread attention in the field of image recognition.
Many interpretability models that have been proposed still have problems of insufficient accuracy and interpretability in medical image disease diagnosis.
We propose feature-driven inference network (FeaInfNet) to solve these problems.
arXiv Detail & Related papers (2023-12-04T13:09:00Z) - Few-Shot Medical Image Segmentation via a Region-enhanced Prototypical
Transformer [20.115149216170327]
Region-enhanced Prototypical Transformer (RPT) is a few-shot learning-based method to mitigate the effects of large intra-class diversity/bias.
By stacking BaT blocks, the proposed RPT can iteratively optimize generated regional prototypes and finally produce rectified and more accurate global prototypes.
arXiv Detail & Related papers (2023-09-09T15:39:38Z) - Multi-task Paired Masking with Alignment Modeling for Medical
Vision-Language Pre-training [55.56609500764344]
We propose a unified framework based on Multi-task Paired Masking with Alignment (MPMA) to integrate the cross-modal alignment task into the joint image-text reconstruction framework.
We also introduce a Memory-Augmented Cross-Modal Fusion (MA-CMF) module to fully integrate visual information to assist report reconstruction.
arXiv Detail & Related papers (2023-05-13T13:53:48Z) - Ambiguous Medical Image Segmentation using Diffusion Models [60.378180265885945]
We introduce a single diffusion model-based approach that produces multiple plausible outputs by learning a distribution over group insights.
Our proposed model generates a distribution of segmentation masks by leveraging the inherent sampling process of diffusion.
Comprehensive results show that our proposed approach outperforms existing state-of-the-art ambiguous segmentation networks.
arXiv Detail & Related papers (2023-04-10T17:58:22Z) - Generalizable Medical Image Segmentation via Random Amplitude Mixup and
Domain-Specific Image Restoration [17.507951655445652]
We present a novel generalizable medical image segmentation method.
To be specific, we design our approach as a multi-task paradigm by combining the segmentation model with a self-supervision domain-specific image restoration module.
We demonstrate the performance of our method on two public generalizable segmentation benchmarks in medical images.
arXiv Detail & Related papers (2022-08-08T03:56:20Z) - Multiple Sclerosis Lesions Segmentation using Attention-Based CNNs in
FLAIR Images [0.2578242050187029]
Multiple Sclerosis (MS) is an autoimmune, and demyelinating disease that leads to lesions in the central nervous system.
Up to now a multitude of multimodality automatic biomedical approaches is used to segment lesions.
Authors propose a method employing just one modality (FLAIR image) to segment MS lesions accurately.
arXiv Detail & Related papers (2022-01-05T21:37:43Z) - Modality Completion via Gaussian Process Prior Variational Autoencoders
for Multi-Modal Glioma Segmentation [75.58395328700821]
We propose a novel model, Multi-modal Gaussian Process Prior Variational Autoencoder (MGP-VAE), to impute one or more missing sub-modalities for a patient scan.
MGP-VAE can leverage the Gaussian Process (GP) prior on the Variational Autoencoder (VAE) to utilize the subjects/patients and sub-modalities correlations.
We show the applicability of MGP-VAE on brain tumor segmentation where either, two, or three of four sub-modalities may be missing.
arXiv Detail & Related papers (2021-07-07T19:06:34Z) - Cross-Modality Brain Tumor Segmentation via Bidirectional
Global-to-Local Unsupervised Domain Adaptation [61.01704175938995]
In this paper, we propose a novel Bidirectional Global-to-Local (BiGL) adaptation framework under a UDA scheme.
Specifically, a bidirectional image synthesis and segmentation module is proposed to segment the brain tumor.
The proposed method outperforms several state-of-the-art unsupervised domain adaptation methods by a large margin.
arXiv Detail & Related papers (2021-05-17T10:11:45Z) - Few-shot Medical Image Segmentation using a Global Correlation Network
with Discriminative Embedding [60.89561661441736]
We propose a novel method for few-shot medical image segmentation.
We construct our few-shot image segmentor using a deep convolutional network trained episodically.
We enhance discriminability of deep embedding to encourage clustering of the feature domains of the same class.
arXiv Detail & Related papers (2020-12-10T04:01:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.