Transferring Ultrahigh-Field Representations for Intensity-Guided Brain
Segmentation of Low-Field Magnetic Resonance Imaging
- URL: http://arxiv.org/abs/2402.08409v1
- Date: Tue, 13 Feb 2024 12:21:06 GMT
- Title: Transferring Ultrahigh-Field Representations for Intensity-Guided Brain
Segmentation of Low-Field Magnetic Resonance Imaging
- Authors: Kwanseok Oh, Jieun Lee, Da-Woon Heo, Dinggang Shen, and Heung-Il Suk
- Abstract summary: The use of 7T MRI is limited by its high cost and lower accessibility compared to low-field (LF) MRI.
This study proposes a deep-learning framework that fuses the input LF magnetic resonance feature representations with the inferred 7T-like feature representations for brain image segmentation tasks.
- Score: 51.92395928517429
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Ultrahigh-field (UHF) magnetic resonance imaging (MRI), i.e., 7T MRI,
provides superior anatomical details of internal brain structures owing to its
enhanced signal-to-noise ratio and susceptibility-induced contrast. However,
the widespread use of 7T MRI is limited by its high cost and lower
accessibility compared to low-field (LF) MRI. This study proposes a
deep-learning framework that systematically fuses the input LF magnetic
resonance feature representations with the inferred 7T-like feature
representations for brain image segmentation tasks in a 7T-absent environment.
Specifically, our adaptive fusion module aggregates 7T-like features derived
from the LF image by a pre-trained network and then refines them to be
effectively assimilable UHF guidance into LF image features. Using
intensity-guided features obtained from such aggregation and assimilation,
segmentation models can recognize subtle structural representations that are
usually difficult to recognize when relying only on LF features. Beyond such
advantages, this strategy can seamlessly be utilized by modulating the contrast
of LF features in alignment with UHF guidance, even when employing arbitrary
segmentation models. Exhaustive experiments demonstrated that the proposed
method significantly outperformed all baseline models on both brain tissue and
whole-brain segmentation tasks; further, it exhibited remarkable adaptability
and scalability by successfully integrating diverse segmentation models and
tasks. These improvements were not only quantifiable but also visible in the
superlative visual quality of segmentation masks.
Related papers
- Unveiling Incomplete Modality Brain Tumor Segmentation: Leveraging Masked Predicted Auto-Encoder and Divergence Learning [6.44069573245889]
Brain tumor segmentation remains a significant challenge, particularly in the context of multi-modal magnetic resonance imaging (MRI)
We propose a novel strategy, which is called masked predicted pre-training, enabling robust feature learning from incomplete modality data.
In the fine-tuning phase, we utilize a knowledge distillation technique to align features between complete and missing modality data, simultaneously enhancing model robustness.
arXiv Detail & Related papers (2024-06-12T20:35:16Z) - MindFormer: Semantic Alignment of Multi-Subject fMRI for Brain Decoding [50.55024115943266]
We introduce a novel semantic alignment method of multi-subject fMRI signals using so-called MindFormer.
This model is specifically designed to generate fMRI-conditioned feature vectors that can be used for conditioning Stable Diffusion model for fMRI- to-image generation or large language model (LLM) for fMRI-to-text generation.
Our experimental results demonstrate that MindFormer generates semantically consistent images and text across different subjects.
arXiv Detail & Related papers (2024-05-28T00:36:25Z) - NeuroPictor: Refining fMRI-to-Image Reconstruction via Multi-individual Pretraining and Multi-level Modulation [55.51412454263856]
This paper proposes to directly modulate the generation process of diffusion models using fMRI signals.
By training with about 67,000 fMRI-image pairs from various individuals, our model enjoys superior fMRI-to-image decoding capacity.
arXiv Detail & Related papers (2024-03-27T02:42:52Z) - Lite-Mind: Towards Efficient and Robust Brain Representation Network [23.132310312815967]
Lite-Mind is a lightweight, efficient, and robust brain representation learning paradigm based on Discrete Frequency Transform (DFT)
We show Lite-Mind achieves an impressive 94.6% fMRI-to-image retrieval accuracy on the NSD dataset for Subject 1, with 98.7% fewer parameters than MindEye.
Lite-Mind is also proven to be able to be migrated to smaller fMRI datasets and establishes a new state-of-the-art for zero-shot classification on the GOD dataset.
arXiv Detail & Related papers (2023-12-06T09:39:38Z) - Compositional Representation Learning for Brain Tumour Segmentation [13.5112749699868]
deep learning models can achieve human expert-level performance given a large amount of data and pixel-level annotations.
We adapt a mixed supervision framework, vMFNet, to learn robust representations using unsupervised learning and weak supervision.
We show that good tumour segmentation performance can be achieved with a large amount of weakly labelled data but only a small amount of fully-annotated data.
arXiv Detail & Related papers (2023-10-10T12:19:39Z) - K-Space-Aware Cross-Modality Score for Synthesized Neuroimage Quality
Assessment [71.27193056354741]
The problem of how to assess cross-modality medical image synthesis has been largely unexplored.
We propose a new metric K-CROSS to spur progress on this challenging problem.
K-CROSS uses a pre-trained multi-modality segmentation network to predict the lesion location.
arXiv Detail & Related papers (2023-07-10T01:26:48Z) - BrainCLIP: Bridging Brain and Visual-Linguistic Representation Via CLIP
for Generic Natural Visual Stimulus Decoding [51.911473457195555]
BrainCLIP is a task-agnostic fMRI-based brain decoding model.
It bridges the modality gap between brain activity, image, and text.
BrainCLIP can reconstruct visual stimuli with high semantic fidelity.
arXiv Detail & Related papers (2023-02-25T03:28:54Z) - Flexible Alignment Super-Resolution Network for Multi-Contrast MRI [7.727046305845654]
Super-Resolution plays a crucial role in preprocessing the low-resolution images for more precise medical analysis.
We propose the Flexible Alignment Super-Resolution Network (FASR-Net) for multi-contrast magnetic resonance images Super-Resolution.
arXiv Detail & Related papers (2022-10-07T11:07:20Z) - Multi-modal Aggregation Network for Fast MR Imaging [85.25000133194762]
We propose a novel Multi-modal Aggregation Network, named MANet, which is capable of discovering complementary representations from a fully sampled auxiliary modality.
In our MANet, the representations from the fully sampled auxiliary and undersampled target modalities are learned independently through a specific network.
Our MANet follows a hybrid domain learning framework, which allows it to simultaneously recover the frequency signal in the $k$-space domain.
arXiv Detail & Related papers (2021-10-15T13:16:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.