Cross-dimensional transfer learning in medical image segmentation with
deep learning
- URL: http://arxiv.org/abs/2307.15872v1
- Date: Sat, 29 Jul 2023 02:50:38 GMT
- Title: Cross-dimensional transfer learning in medical image segmentation with
deep learning
- Authors: Hicham Messaoudi, Ahror Belaid, Douraied Ben Salem, Pierre-Henri Conze
- Abstract summary: We introduce an efficient way to transfer the efficiency of a 2D classification network trained on natural images to 2D, 3D uni- and multi-modal medical image segmentation applications.
In this paper, we designed novel architectures based on two key principles: weight transfer by embedding a 2D pre-trained encoder into a higher dimensional U-Net, and dimensional transfer by expanding a 2D segmentation network into a higher dimension one.
- Score: 0.4588028371034407
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Over the last decade, convolutional neural networks have emerged and advanced
the state-of-the-art in various image analysis and computer vision
applications. The performance of 2D image classification networks is constantly
improving and being trained on databases made of millions of natural images.
However, progress in medical image analysis has been hindered by limited
annotated data and acquisition constraints. These limitations are even more
pronounced given the volumetry of medical imaging data. In this paper, we
introduce an efficient way to transfer the efficiency of a 2D classification
network trained on natural images to 2D, 3D uni- and multi-modal medical image
segmentation applications. In this direction, we designed novel architectures
based on two key principles: weight transfer by embedding a 2D pre-trained
encoder into a higher dimensional U-Net, and dimensional transfer by expanding
a 2D segmentation network into a higher dimension one. The proposed networks
were tested on benchmarks comprising different modalities: MR, CT, and
ultrasound images. Our 2D network ranked first on the CAMUS challenge dedicated
to echo-cardiographic data segmentation and surpassed the state-of-the-art.
Regarding 2D/3D MR and CT abdominal images from the CHAOS challenge, our
approach largely outperformed the other 2D-based methods described in the
challenge paper on Dice, RAVD, ASSD, and MSSD scores and ranked third on the
online evaluation platform. Our 3D network applied to the BraTS 2022
competition also achieved promising results, reaching an average Dice score of
91.69% (91.22%) for the whole tumor, 83.23% (84.77%) for the tumor core, and
81.75% (83.88%) for enhanced tumor using the approach based on weight
(dimensional) transfer. Experimental and qualitative results illustrate the
effectiveness of our methods for multi-dimensional medical image segmentation.
Related papers
- Deep Convolutional Neural Networks on Multiclass Classification of Three-Dimensional Brain Images for Parkinson's Disease Stage Prediction [2.931680194227131]
We developed a model capable of accurately predicting Parkinson's disease stages.
We used the entire three-dimensional (3D) brain images as input.
We incorporated an attention mechanism to account for the varying importance of different slices in the prediction process.
arXiv Detail & Related papers (2024-10-31T05:40:08Z) - 3D-CT-GPT: Generating 3D Radiology Reports through Integration of Large Vision-Language Models [51.855377054763345]
This paper introduces 3D-CT-GPT, a Visual Question Answering (VQA)-based medical visual language model for generating radiology reports from 3D CT scans.
Experiments on both public and private datasets demonstrate that 3D-CT-GPT significantly outperforms existing methods in terms of report accuracy and quality.
arXiv Detail & Related papers (2024-09-28T12:31:07Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - Improving 3D Imaging with Pre-Trained Perpendicular 2D Diffusion Models [52.529394863331326]
We propose a novel approach using two perpendicular pre-trained 2D diffusion models to solve the 3D inverse problem.
Our method is highly effective for 3D medical image reconstruction tasks, including MRI Z-axis super-resolution, compressed sensing MRI, and sparse-view CT.
arXiv Detail & Related papers (2023-03-15T08:28:06Z) - Med-DANet: Dynamic Architecture Network for Efficient Medical Volumetric
Segmentation [13.158995287578316]
We propose a dynamic architecture network named Med-DANet to achieve effective accuracy and efficiency trade-off.
For each slice of the input 3D MRI volume, our proposed method learns a slice-specific decision by the Decision Network.
Our proposed method achieves comparable or better results than previous state-of-the-art methods for 3D MRI brain tumor segmentation.
arXiv Detail & Related papers (2022-06-14T03:25:58Z) - Harmonizing Pathological and Normal Pixels for Pseudo-healthy Synthesis [68.5287824124996]
We present a new type of discriminator, the segmentor, to accurately locate the lesions and improve the visual quality of pseudo-healthy images.
We apply the generated images into medical image enhancement and utilize the enhanced results to cope with the low contrast problem.
Comprehensive experiments on the T2 modality of BraTS demonstrate that the proposed method substantially outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2022-03-29T08:41:17Z) - Simultaneous Alignment and Surface Regression Using Hybrid 2D-3D
Networks for 3D Coherent Layer Segmentation of Retina OCT Images [33.99874168018807]
In this study, a novel framework based on hybrid 2D-3D convolutional neural networks (CNNs) is proposed to obtain continuous 3D retinal layer surfaces from OCT.
Our framework achieves superior results to state-of-the-art 2D methods in terms of both layer segmentation accuracy and cross-B-scan 3D continuity.
arXiv Detail & Related papers (2022-03-04T15:55:09Z) - Mutual Attention-based Hybrid Dimensional Network for Multimodal Imaging
Computer-aided Diagnosis [4.657804635843888]
We propose a novel mutual attention-based hybrid dimensional network for MultiModal 3D medical image classification (MMNet)
The hybrid dimensional network integrates 2D CNN with 3D convolution modules to generate deeper and more informative feature maps.
We further design a mutual attention framework in the network to build the region-wise consistency in similar stereoscopic regions of different image modalities.
arXiv Detail & Related papers (2022-01-24T02:31:25Z) - Efficient embedding network for 3D brain tumor segmentation [0.33727511459109777]
In this paper, we investigate a way to transfer the performance of a two-dimensional classiffication network for the purpose of three-dimensional semantic segmentation of brain tumors.
As the input data is in 3D, the first layers of the encoder are devoted to the reduction of the third dimension in order to fit the input of the EfficientNet network.
Experimental results on validation and test data from the BraTS 2020 challenge demonstrate that the proposed method achieve promising performance.
arXiv Detail & Related papers (2020-11-22T16:17:29Z) - Volumetric Medical Image Segmentation: A 3D Deep Coarse-to-fine
Framework and Its Adversarial Examples [74.92488215859991]
We propose a novel 3D-based coarse-to-fine framework to efficiently tackle these challenges.
The proposed 3D-based framework outperforms their 2D counterparts by a large margin since it can leverage the rich spatial information along all three axes.
We conduct experiments on three datasets, the NIH pancreas dataset, the JHMI pancreas dataset and the JHMI pathological cyst dataset.
arXiv Detail & Related papers (2020-10-29T15:39:19Z) - 2.75D: Boosting learning by representing 3D Medical imaging to 2D
features for small data [54.223614679807994]
3D convolutional neural networks (CNNs) have started to show superior performance to 2D CNNs in numerous deep learning tasks.
Applying transfer learning on 3D CNN is challenging due to a lack of publicly available pre-trained 3D models.
In this work, we proposed a novel 2D strategical representation of volumetric data, namely 2.75D.
As a result, 2D CNN networks can also be used to learn volumetric information.
arXiv Detail & Related papers (2020-02-11T08:24:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.