Medical Transformer: Universal Brain Encoder for 3D MRI Analysis
- URL: http://arxiv.org/abs/2104.13633v1
- Date: Wed, 28 Apr 2021 08:34:21 GMT
- Title: Medical Transformer: Universal Brain Encoder for 3D MRI Analysis
- Authors: Eunji Jun, Seungwoo Jeong, Da-Woon Heo, Heung-Il Suk
- Abstract summary: Existing 3D-based methods have transferred the pre-trained models to downstream tasks.
They demand a massive amount of parameters to train the model for 3D medical imaging.
We propose a novel transfer learning framework, called Medical Transformer, that effectively models 3D volumetric images in the form of a sequence of 2D image slices.
- Score: 1.6287500717172143
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Transfer learning has gained attention in medical image analysis due to
limited annotated 3D medical datasets for training data-driven deep learning
models in the real world. Existing 3D-based methods have transferred the
pre-trained models to downstream tasks, which achieved promising results with
only a small number of training samples. However, they demand a massive amount
of parameters to train the model for 3D medical imaging. In this work, we
propose a novel transfer learning framework, called Medical Transformer, that
effectively models 3D volumetric images in the form of a sequence of 2D image
slices. To make a high-level representation in 3D-form empowering spatial
relations better, we take a multi-view approach that leverages plenty of
information from the three planes of 3D volume, while providing
parameter-efficient training. For building a source model generally applicable
to various tasks, we pre-train the model in a self-supervised learning manner
for masked encoding vector prediction as a proxy task, using a large-scale
normal, healthy brain magnetic resonance imaging (MRI) dataset. Our pre-trained
model is evaluated on three downstream tasks: (i) brain disease diagnosis, (ii)
brain age prediction, and (iii) brain tumor segmentation, which are actively
studied in brain MRI research. The experimental results show that our Medical
Transformer outperforms the state-of-the-art transfer learning methods,
efficiently reducing the number of parameters up to about 92% for
classification and
Related papers
- Domain Aware Multi-Task Pretraining of 3D Swin Transformer for T1-weighted Brain MRI [4.453300553789746]
We propose novel domain-aware multi-task learning tasks to pretrain a 3D Swin Transformer for brain magnetic resonance imaging (MRI)
Our method considers the domain knowledge in brain MRI by incorporating brain anatomy and morphology as well as standard pretext tasks adapted for 3D imaging in a contrastive learning setting.
Our method outperforms existing supervised and self-supervised methods in three downstream tasks of Alzheimer's disease classification, Parkinson's disease classification, and age prediction tasks.
arXiv Detail & Related papers (2024-10-01T05:21:02Z) - 3D-CT-GPT: Generating 3D Radiology Reports through Integration of Large Vision-Language Models [51.855377054763345]
This paper introduces 3D-CT-GPT, a Visual Question Answering (VQA)-based medical visual language model for generating radiology reports from 3D CT scans.
Experiments on both public and private datasets demonstrate that 3D-CT-GPT significantly outperforms existing methods in terms of report accuracy and quality.
arXiv Detail & Related papers (2024-09-28T12:31:07Z) - Brain3D: Generating 3D Objects from fMRI [76.41771117405973]
We design a novel 3D object representation learning method, Brain3D, that takes as input the fMRI data of a subject.
We show that our model captures the distinct functionalities of each region of human vision system.
Preliminary evaluations indicate that Brain3D can successfully identify the disordered brain regions in simulated scenarios.
arXiv Detail & Related papers (2024-05-24T06:06:11Z) - Generative Enhancement for 3D Medical Images [74.17066529847546]
We propose GEM-3D, a novel generative approach to the synthesis of 3D medical images.
Our method begins with a 2D slice, noted as the informed slice to serve the patient prior, and propagates the generation process using a 3D segmentation mask.
By decomposing the 3D medical images into masks and patient prior information, GEM-3D offers a flexible yet effective solution for generating versatile 3D images.
arXiv Detail & Related papers (2024-03-19T15:57:04Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - Adapting Pre-trained Vision Transformers from 2D to 3D through Weight
Inflation Improves Medical Image Segmentation [19.693778706169752]
We use a weight inflation strategy to adapt pre-trained Transformers from 2D to 3D, retaining the benefit of both transfer learning and depth information.
Our approach achieves state-of-the-art performances across a broad range of 3D medical image datasets.
arXiv Detail & Related papers (2023-02-08T19:38:13Z) - Attentive Symmetric Autoencoder for Brain MRI Segmentation [56.02577247523737]
We propose a novel Attentive Symmetric Auto-encoder based on Vision Transformer (ViT) for 3D brain MRI segmentation tasks.
In the pre-training stage, the proposed auto-encoder pays more attention to reconstruct the informative patches according to the gradient metrics.
Experimental results show that our proposed attentive symmetric auto-encoder outperforms the state-of-the-art self-supervised learning methods and medical image segmentation models.
arXiv Detail & Related papers (2022-09-19T09:43:19Z) - Self-Supervised Learning for 3D Medical Image Analysis using 3D SimCLR
and Monte Carlo Dropout [4.190672342471606]
We propose a 3D self-supervised method that is based on the contrastive (SimCLR) method.
We show that employing Bayesian neural networks during the inference phase can further enhance the results on the downstream tasks.
arXiv Detail & Related papers (2021-09-29T09:15:41Z) - Planar 3D Transfer Learning for End to End Unimodal MRI Unbalanced Data
Segmentation [0.0]
We present a novel approach of 2D to 3D transfer learning based on mapping pre-trained 2D convolutional neural network weights into planar 3D kernels.
The method is validated by the proposed planar 3D res-u-net network with encoder transferred from the 2D VGG-16.
arXiv Detail & Related papers (2020-11-23T17:11:50Z) - Modelling the Distribution of 3D Brain MRI using a 2D Slice VAE [66.63629641650572]
We propose a method to model 3D MR brain volumes distribution by combining a 2D slice VAE with a Gaussian model that captures the relationships between slices.
We also introduce a novel evaluation method for generated volumes that quantifies how well their segmentations match those of true brain anatomy.
arXiv Detail & Related papers (2020-07-09T13:23:15Z) - 3D Self-Supervised Methods for Medical Imaging [7.65168530693281]
We propose 3D versions for five different self-supervised methods, in the form of proxy tasks.
Our methods facilitate neural network feature learning from unlabeled 3D images, aiming to reduce the required cost for expert annotation.
The developed algorithms are 3D Contrastive Predictive Coding, 3D Rotation prediction, 3D Jigsaw puzzles, Relative 3D patch location, and 3D Exemplar networks.
arXiv Detail & Related papers (2020-06-06T09:56:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.