3D Self-Supervised Methods for Medical Imaging
- URL: http://arxiv.org/abs/2006.03829v3
- Date: Mon, 2 Nov 2020 10:43:10 GMT
- Title: 3D Self-Supervised Methods for Medical Imaging
- Authors: Aiham Taleb, Winfried Loetzsch, Noel Danz, Julius Severin, Thomas
Gaertner, Benjamin Bergner, and Christoph Lippert
- Abstract summary: We propose 3D versions for five different self-supervised methods, in the form of proxy tasks.
Our methods facilitate neural network feature learning from unlabeled 3D images, aiming to reduce the required cost for expert annotation.
The developed algorithms are 3D Contrastive Predictive Coding, 3D Rotation prediction, 3D Jigsaw puzzles, Relative 3D patch location, and 3D Exemplar networks.
- Score: 7.65168530693281
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Self-supervised learning methods have witnessed a recent surge of interest
after proving successful in multiple application fields. In this work, we
leverage these techniques, and we propose 3D versions for five different
self-supervised methods, in the form of proxy tasks. Our methods facilitate
neural network feature learning from unlabeled 3D images, aiming to reduce the
required cost for expert annotation. The developed algorithms are 3D
Contrastive Predictive Coding, 3D Rotation prediction, 3D Jigsaw puzzles,
Relative 3D patch location, and 3D Exemplar networks. Our experiments show that
pretraining models with our 3D tasks yields more powerful semantic
representations, and enables solving downstream tasks more accurately and
efficiently, compared to training the models from scratch and to pretraining
them on 2D slices. We demonstrate the effectiveness of our methods on three
downstream tasks from the medical imaging domain: i) Brain Tumor Segmentation
from 3D MRI, ii) Pancreas Tumor Segmentation from 3D CT, and iii) Diabetic
Retinopathy Detection from 2D Fundus images. In each task, we assess the gains
in data-efficiency, performance, and speed of convergence. Interestingly, we
also find gains when transferring the learned representations, by our methods,
from a large unlabeled 3D corpus to a small downstream-specific dataset. We
achieve results competitive to state-of-the-art solutions at a fraction of the
computational expense. We publish our implementations for the developed
algorithms (both 3D and 2D versions) as an open-source library, in an effort to
allow other researchers to apply and extend our methods on their datasets.
Related papers
- OV-Uni3DETR: Towards Unified Open-Vocabulary 3D Object Detection via Cycle-Modality Propagation [67.56268991234371]
OV-Uni3DETR achieves the state-of-the-art performance on various scenarios, surpassing existing methods by more than 6% on average.
Code and pre-trained models will be released later.
arXiv Detail & Related papers (2024-03-28T17:05:04Z) - PonderV2: Pave the Way for 3D Foundation Model with A Universal
Pre-training Paradigm [114.47216525866435]
We introduce a novel universal 3D pre-training framework designed to facilitate the acquisition of efficient 3D representation.
For the first time, PonderV2 achieves state-of-the-art performance on 11 indoor and outdoor benchmarks, implying its effectiveness.
arXiv Detail & Related papers (2023-10-12T17:59:57Z) - Joint Embedding of 2D and 3D Networks for Medical Image Anomaly
Detection [0.8122270502556374]
We develop a method for combining the strength of the 3D network and the strength of the 2D network through joint embedding.
We show that the proposed method achieves better performance in both classification and segmentation tasks compared to the SoTA method.
arXiv Detail & Related papers (2022-12-21T11:28:52Z) - Super Images -- A New 2D Perspective on 3D Medical Imaging Analysis [0.0]
We present a simple yet effective 2D method to handle 3D data while efficiently embedding the 3D knowledge during training.
Our method generates a super-resolution image by stitching slices side by side in the 3D image.
While attaining equal, if not superior, results to 3D networks utilizing only 2D counterparts, the model complexity is reduced by around threefold.
arXiv Detail & Related papers (2022-05-05T09:59:03Z) - Homography Loss for Monocular 3D Object Detection [54.04870007473932]
A differentiable loss function, termed as Homography Loss, is proposed to achieve the goal, which exploits both 2D and 3D information.
Our method yields the best performance compared with the other state-of-the-arts by a large margin on KITTI 3D datasets.
arXiv Detail & Related papers (2022-04-02T03:48:03Z) - Self-Supervised Learning for 3D Medical Image Analysis using 3D SimCLR
and Monte Carlo Dropout [4.190672342471606]
We propose a 3D self-supervised method that is based on the contrastive (SimCLR) method.
We show that employing Bayesian neural networks during the inference phase can further enhance the results on the downstream tasks.
arXiv Detail & Related papers (2021-09-29T09:15:41Z) - 3-Dimensional Deep Learning with Spatial Erasing for Unsupervised
Anomaly Segmentation in Brain MRI [55.97060983868787]
We investigate whether using increased spatial context by using MRI volumes combined with spatial erasing leads to improved unsupervised anomaly segmentation performance.
We compare 2D variational autoencoder (VAE) to their 3D counterpart, propose 3D input erasing, and systemically study the impact of the data set size on the performance.
Our best performing 3D VAE with input erasing leads to an average DICE score of 31.40% compared to 25.76% for the 2D VAE.
arXiv Detail & Related papers (2021-09-14T09:17:27Z) - Medical Transformer: Universal Brain Encoder for 3D MRI Analysis [1.6287500717172143]
Existing 3D-based methods have transferred the pre-trained models to downstream tasks.
They demand a massive amount of parameters to train the model for 3D medical imaging.
We propose a novel transfer learning framework, called Medical Transformer, that effectively models 3D volumetric images in the form of a sequence of 2D image slices.
arXiv Detail & Related papers (2021-04-28T08:34:21Z) - Revisiting 3D Context Modeling with Supervised Pre-training for
Universal Lesion Detection in CT Slices [48.85784310158493]
We propose a Modified Pseudo-3D Feature Pyramid Network (MP3D FPN) to efficiently extract 3D context enhanced 2D features for universal lesion detection in CT slices.
With the novel pre-training method, the proposed MP3D FPN achieves state-of-the-art detection performance on the DeepLesion dataset.
The proposed 3D pre-trained weights can potentially be used to boost the performance of other 3D medical image analysis tasks.
arXiv Detail & Related papers (2020-12-16T07:11:16Z) - Planar 3D Transfer Learning for End to End Unimodal MRI Unbalanced Data
Segmentation [0.0]
We present a novel approach of 2D to 3D transfer learning based on mapping pre-trained 2D convolutional neural network weights into planar 3D kernels.
The method is validated by the proposed planar 3D res-u-net network with encoder transferred from the 2D VGG-16.
arXiv Detail & Related papers (2020-11-23T17:11:50Z) - 2.75D: Boosting learning by representing 3D Medical imaging to 2D
features for small data [54.223614679807994]
3D convolutional neural networks (CNNs) have started to show superior performance to 2D CNNs in numerous deep learning tasks.
Applying transfer learning on 3D CNN is challenging due to a lack of publicly available pre-trained 3D models.
In this work, we proposed a novel 2D strategical representation of volumetric data, namely 2.75D.
As a result, 2D CNN networks can also be used to learn volumetric information.
arXiv Detail & Related papers (2020-02-11T08:24:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.