Multi-view Cross-Modality MR Image Translation for Vestibular Schwannoma
and Cochlea Segmentation
- URL: http://arxiv.org/abs/2303.14998v1
- Date: Mon, 27 Mar 2023 08:42:10 GMT
- Title: Multi-view Cross-Modality MR Image Translation for Vestibular Schwannoma
and Cochlea Segmentation
- Authors: Bogyeong Kang, Hyeonyeong Nam, Ji-Wung Han, Keun-Soo Heo, and Tae-Eui
Kam
- Abstract summary: We propose a multi-view image translation framework, which can translate contrast-enhanced T1 (ceT1) MR imaging to high-resolution T2 (hrT2) MR imaging.
We adopt two image translation models in parallel that use a pixel-level consistent constraint and a patch-level contrastive constraint, respectively.
Thereby, we can augment pseudo-hrT2 images reflecting different perspectives, which eventually lead to a high-performing segmentation model.
- Score: 0.7829352305480285
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we propose a multi-view image translation framework, which can
translate contrast-enhanced T1 (ceT1) MR imaging to high-resolution T2 (hrT2)
MR imaging for unsupervised vestibular schwannoma and cochlea segmentation. We
adopt two image translation models in parallel that use a pixel-level
consistent constraint and a patch-level contrastive constraint, respectively.
Thereby, we can augment pseudo-hrT2 images reflecting different perspectives,
which eventually lead to a high-performing segmentation model. Our experimental
results on the CrossMoDA challenge show that the proposed method achieved
enhanced performance on the vestibular schwannoma and cochlea segmentation.
Related papers
- Anatomical Conditioning for Contrastive Unpaired Image-to-Image Translation of Optical Coherence Tomography Images [0.0]
We study the problem employing an optical coherence tomography ( OCT) data set of Spectralis- OCT and Home- OCT images.
I2I translation is challenging because the images are unpaired.
Our approach increases the similarity between the style-translated images and the target distribution.
arXiv Detail & Related papers (2024-04-08T11:20:28Z) - Referee Can Play: An Alternative Approach to Conditional Generation via
Model Inversion [35.21106030549071]
Diffusion Probabilistic Models (DPMs) are dominant force in text-to-image generation tasks.
We propose an alternative view of state-of-the-art DPMs as a way of inverting advanced Vision-Language Models (VLMs)
By directly optimizing images with the supervision of discriminative VLMs, the proposed method can potentially achieve a better text-image alignment.
arXiv Detail & Related papers (2024-02-26T05:08:40Z) - Dual-scale Enhanced and Cross-generative Consistency Learning for Semi-supervised Medical Image Segmentation [49.57907601086494]
Medical image segmentation plays a crucial role in computer-aided diagnosis.
We propose a novel Dual-scale Enhanced and Cross-generative consistency learning framework for semi-supervised medical image (DEC-Seg)
arXiv Detail & Related papers (2023-12-26T12:56:31Z) - Spatial and Modal Optimal Transport for Fast Cross-Modal MRI Reconstruction [54.19448988321891]
We propose an end-to-end deep learning framework that utilizes T1-weighted images (T1WIs) as auxiliary modalities to expedite T2WIs' acquisitions.
We employ Optimal Transport (OT) to synthesize T2WIs by aligning T1WIs and performing cross-modal synthesis.
We prove that the reconstructed T2WIs and the synthetic T2WIs become closer on the T2 image manifold with iterations increasing.
arXiv Detail & Related papers (2023-05-04T12:20:51Z) - BBDM: Image-to-image Translation with Brownian Bridge Diffusion Models [50.39417112077254]
A novel image-to-image translation method based on the Brownian Bridge Diffusion Model (BBDM) is proposed.
To the best of our knowledge, it is the first work that proposes Brownian Bridge diffusion process for image-to-image translation.
arXiv Detail & Related papers (2022-05-16T13:47:02Z) - Cross-View Panorama Image Synthesis [68.35351563852335]
PanoGAN is a novel adversarial feedback GAN framework named.
PanoGAN enables high-quality panorama image generation with more convincing details than state-of-the-art approaches.
arXiv Detail & Related papers (2022-03-22T15:59:44Z) - An Efficient Smoothing and Thresholding Image Segmentation Framework
with Weighted Anisotropic-Isotropic Total Variation [1.9581049654950413]
We present a multi-stage image segmentation framework that incorporates a weighted difference of anisotropic isotropic variation (AITV)
In the second stage, we threshold the smoothed image by $K$-meansizer to obtain the final result.
arXiv Detail & Related papers (2022-02-21T10:57:16Z) - Multi-domain Unsupervised Image-to-Image Translation with Appearance
Adaptive Convolution [62.4972011636884]
We propose a novel multi-domain unsupervised image-to-image translation (MDUIT) framework.
We exploit the decomposed content feature and appearance adaptive convolution to translate an image into a target appearance.
We show that the proposed method produces visually diverse and plausible results in multiple domains compared to the state-of-the-art methods.
arXiv Detail & Related papers (2022-02-06T14:12:34Z) - Smoothing the Disentangled Latent Style Space for Unsupervised
Image-to-Image Translation [56.55178339375146]
Image-to-Image (I2I) multi-domain translation models are usually evaluated also using the quality of their semantic results.
We propose a new training protocol based on three specific losses which help a translation network to learn a smooth and disentangled latent style space.
arXiv Detail & Related papers (2021-06-16T17:58:21Z) - Flow-based Deformation Guidance for Unpaired Multi-Contrast MRI
Image-to-Image Translation [7.8333615755210175]
In this paper, we introduce a novel approach to unpaired image-to-image translation based on the invertible architecture.
We utilize the temporal information between consecutive slices to provide more constraints to the optimization for transforming one domain to another in unpaired medical images.
arXiv Detail & Related papers (2020-12-03T09:10:22Z) - Unified cross-modality feature disentangler for unsupervised
multi-domain MRI abdomen organs segmentation [3.3504365823045044]
Our contribution is a unified cross-modality feature disentagling approach for multi-domain image translation and multiple organ segmentation.
Using CT as the labeled source domain, our approach learns to segment multi-modal (T1-weighted and T2-weighted) MRI having no labeled data.
Our approach produced an average Dice similarity coefficient (DSC) of 0.85 for T1w and 0.90 for T2w MRI for multi-organ segmentation.
arXiv Detail & Related papers (2020-07-19T13:33:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.