Representation Disentanglement for Multi-modal MR Analysis
- URL: http://arxiv.org/abs/2102.11456v1
- Date: Tue, 23 Feb 2021 02:08:38 GMT
- Title: Representation Disentanglement for Multi-modal MR Analysis
- Authors: Jiahong Ouyang, Ehsan Adeli, Kilian M Pohl, Qingyu Zhao, Greg
Zaharchuk
- Abstract summary: Recent works have suggested that multi-modal deep learning analysis can benefit from explicitly disentangling anatomical (shape) and modality (appearance) representations from the images.
We propose a margin loss that regularizes the similarity relationships of the representations across subjects and modalities.
To enable a robust training, we introduce a modified conditional convolution to design a single model for encoding images of all modalities.
- Score: 15.498244253687337
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-modal MR images are widely used in neuroimaging applications to provide
complementary information about the brain structures. Recent works have
suggested that multi-modal deep learning analysis can benefit from explicitly
disentangling anatomical (shape) and modality (appearance) representations from
the images. In this work, we challenge existing strategies by showing that they
do not naturally lead to representation disentanglement both in theory and in
practice. To address this issue, we propose a margin loss that regularizes the
similarity relationships of the representations across subjects and modalities.
To enable a robust training, we further introduce a modified conditional
convolution to design a single model for encoding images of all modalities.
Lastly, we propose a fusion function to combine the disentangled anatomical
representations as a set of modality-invariant features for downstream tasks.
We evaluate the proposed method on three multi-modal neuroimaging datasets.
Experiments show that our proposed method can achieve superior disentangled
representations compared to existing disentanglement strategies. Results also
indicate that the fused anatomical representation has great potential in the
downstream task of zero-dose PET reconstruction and brain tumor segmentation.
Related papers
- MedMAP: Promoting Incomplete Multi-modal Brain Tumor Segmentation with Alignment [20.358300924109162]
In clinical practice, certain modalities of MRI may be missing, which presents a more difficult scenario.
Knowledge Distillation, Domain Adaption, and Shared Latent Space have emerged as commonly promising strategies.
We propose a novel paradigm that aligns latent features of involved modalities to a well-defined distribution anchor as the substitution of the pre-trained model.
arXiv Detail & Related papers (2024-08-18T13:16:30Z) - Disentangled Multimodal Brain MR Image Translation via Transformer-based
Modality Infuser [12.402947207350394]
We propose a transformer-based modality infuser designed to synthesize multimodal brain MR images.
In our method, we extract modality-agnostic features from the encoder and then transform them into modality-specific features.
We carried out experiments on the BraTS 2018 dataset, translating between four MR modalities.
arXiv Detail & Related papers (2024-02-01T06:34:35Z) - Patched Diffusion Models for Unsupervised Anomaly Detection in Brain MRI [55.78588835407174]
We propose a method that reformulates the generation task of diffusion models as a patch-based estimation of healthy brain anatomy.
We evaluate our approach on data of tumors and multiple sclerosis lesions and demonstrate a relative improvement of 25.1% compared to existing baselines.
arXiv Detail & Related papers (2023-03-07T09:40:22Z) - Model-Guided Multi-Contrast Deep Unfolding Network for MRI
Super-resolution Reconstruction [68.80715727288514]
We show how to unfold an iterative MGDUN algorithm into a novel model-guided deep unfolding network by taking the MRI observation matrix.
In this paper, we propose a novel Model-Guided interpretable Deep Unfolding Network (MGDUN) for medical image SR reconstruction.
arXiv Detail & Related papers (2022-09-15T03:58:30Z) - Learning Multi-Modal Brain Tumor Segmentation from Privileged
Semi-Paired MRI Images with Curriculum Disentanglement Learning [4.43142018105102]
We present a novel two-step (intra-modality and inter-modality) curriculum disentanglement learning framework for brain tumor segmentation.
In the first step, we propose to conduct reconstruction and segmentation with augmented intra-modality style-consistent images.
In the second step, the model jointly performs reconstruction, unsupervised/supervised translation, and segmentation for both unpaired and paired inter-modality images.
arXiv Detail & Related papers (2022-08-26T16:52:43Z) - Multi-modal Aggregation Network for Fast MR Imaging [85.25000133194762]
We propose a novel Multi-modal Aggregation Network, named MANet, which is capable of discovering complementary representations from a fully sampled auxiliary modality.
In our MANet, the representations from the fully sampled auxiliary and undersampled target modalities are learned independently through a specific network.
Our MANet follows a hybrid domain learning framework, which allows it to simultaneously recover the frequency signal in the $k$-space domain.
arXiv Detail & Related papers (2021-10-15T13:16:59Z) - Modality Completion via Gaussian Process Prior Variational Autoencoders
for Multi-Modal Glioma Segmentation [75.58395328700821]
We propose a novel model, Multi-modal Gaussian Process Prior Variational Autoencoder (MGP-VAE), to impute one or more missing sub-modalities for a patient scan.
MGP-VAE can leverage the Gaussian Process (GP) prior on the Variational Autoencoder (VAE) to utilize the subjects/patients and sub-modalities correlations.
We show the applicability of MGP-VAE on brain tumor segmentation where either, two, or three of four sub-modalities may be missing.
arXiv Detail & Related papers (2021-07-07T19:06:34Z) - Self-Attentive Spatial Adaptive Normalization for Cross-Modality Domain
Adaptation [9.659642285903418]
Cross-modality synthesis of medical images to reduce the costly annotation burden by radiologists.
We present a novel approach for image-to-image translation in medical images, capable of supervised or unsupervised (unpaired image data) setups.
arXiv Detail & Related papers (2021-03-05T16:22:31Z) - Deep Co-Attention Network for Multi-View Subspace Learning [73.3450258002607]
We propose a deep co-attention network for multi-view subspace learning.
It aims to extract both the common information and the complementary information in an adversarial setting.
In particular, it uses a novel cross reconstruction loss and leverages the label information to guide the construction of the latent representation.
arXiv Detail & Related papers (2021-02-15T18:46:44Z) - Image Response Regression via Deep Neural Networks [4.646077947295938]
We propose a novel nonparametric approach in the framework of spatially varying coefficient models, where the spatially varying functions are estimated through deep neural networks.
A key idea in our approach is to treat the image voxels as spatial effective samples, which alleviates the limited sample size issue that haunts the majority of medical imaging studies.
We demonstrate the efficacy of the method through intensive simulations, and further illustrate its advantages analyses of two functional magnetic resonance imaging datasets.
arXiv Detail & Related papers (2020-06-17T14:45:26Z) - Hi-Net: Hybrid-fusion Network for Multi-modal MR Image Synthesis [143.55901940771568]
We propose a novel Hybrid-fusion Network (Hi-Net) for multi-modal MR image synthesis.
In our Hi-Net, a modality-specific network is utilized to learn representations for each individual modality.
A multi-modal synthesis network is designed to densely combine the latent representation with hierarchical features from each modality.
arXiv Detail & Related papers (2020-02-11T08:26:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.