Advancing 3D Medical Image Analysis with Variable Dimension Transform
based Supervised 3D Pre-training
- URL: http://arxiv.org/abs/2201.01426v1
- Date: Wed, 5 Jan 2022 03:11:21 GMT
- Title: Advancing 3D Medical Image Analysis with Variable Dimension Transform
based Supervised 3D Pre-training
- Authors: Shu Zhang, Zihao Li, Hong-Yu Zhou, Jiechao Ma, Yizhou Yu
- Abstract summary: This paper revisits an innovative yet simple fully-supervised 3D network pre-training framework.
With a redesigned 3D network architecture, reformulated natural images are used to address the problem of data scarcity.
Comprehensive experiments on four benchmark datasets demonstrate that the proposed pre-trained models can effectively accelerate convergence.
- Score: 45.90045513731704
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The difficulties in both data acquisition and annotation substantially
restrict the sample sizes of training datasets for 3D medical imaging
applications. As a result, constructing high-performance 3D convolutional
neural networks from scratch remains a difficult task in the absence of a
sufficient pre-training parameter. Previous efforts on 3D pre-training have
frequently relied on self-supervised approaches, which use either predictive or
contrastive learning on unlabeled data to build invariant 3D representations.
However, because of the unavailability of large-scale supervision information,
obtaining semantically invariant and discriminative representations from these
learning frameworks remains problematic. In this paper, we revisit an
innovative yet simple fully-supervised 3D network pre-training framework to
take advantage of semantic supervisions from large-scale 2D natural image
datasets. With a redesigned 3D network architecture, reformulated natural
images are used to address the problem of data scarcity and develop powerful 3D
representations. Comprehensive experiments on four benchmark datasets
demonstrate that the proposed pre-trained models can effectively accelerate
convergence while also improving accuracy for a variety of 3D medical imaging
tasks such as classification, segmentation and detection. In addition, as
compared to training from scratch, it can save up to 60% of annotation efforts.
On the NIH DeepLesion dataset, it likewise achieves state-of-the-art detection
performance, outperforming earlier self-supervised and fully-supervised
pre-training approaches, as well as methods that do training from scratch. To
facilitate further development of 3D medical models, our code and pre-trained
model weights are publicly available at https://github.com/urmagicsmine/CSPR.
Related papers
- Cross-D Conv: Cross-Dimensional Transferable Knowledge Base via Fourier Shifting Operation [3.69758875412828]
Cross-D Conv operation bridges the dimensional gap by learning the phase shifting in the Fourier domain.
Our method enables seamless weight transfer between 2D and 3D convolution operations, effectively facilitating cross-dimensional learning.
arXiv Detail & Related papers (2024-11-02T13:03:44Z) - Enhancing Generalizability of Representation Learning for Data-Efficient 3D Scene Understanding [50.448520056844885]
We propose a generative Bayesian network to produce diverse synthetic scenes with real-world patterns.
A series of experiments robustly display our method's consistent superiority over existing state-of-the-art pre-training approaches.
arXiv Detail & Related papers (2024-06-17T07:43:53Z) - FILP-3D: Enhancing 3D Few-shot Class-incremental Learning with
Pre-trained Vision-Language Models [62.663113296987085]
Few-shot class-incremental learning aims to mitigate the catastrophic forgetting issue when a model is incrementally trained on limited data.
We introduce two novel components: the Redundant Feature Eliminator (RFE) and the Spatial Noise Compensator (SNC)
Considering the imbalance in existing 3D datasets, we also propose new evaluation metrics that offer a more nuanced assessment of a 3D FSCIL model.
arXiv Detail & Related papers (2023-12-28T14:52:07Z) - 3D Adversarial Augmentations for Robust Out-of-Domain Predictions [115.74319739738571]
We focus on improving the generalization to out-of-domain data.
We learn a set of vectors that deform the objects in an adversarial fashion.
We perform adversarial augmentation by applying the learned sample-independent vectors to the available objects when training a model.
arXiv Detail & Related papers (2023-08-29T17:58:55Z) - Video Pretraining Advances 3D Deep Learning on Chest CT Tasks [63.879848037679224]
Pretraining on large natural image classification datasets has aided model development on data-scarce 2D medical tasks.
These 2D models have been surpassed by 3D models on 3D computer vision benchmarks.
We show video pretraining for 3D models can enable higher performance on smaller datasets for 3D medical tasks.
arXiv Detail & Related papers (2023-04-02T14:46:58Z) - Super Images -- A New 2D Perspective on 3D Medical Imaging Analysis [0.0]
We present a simple yet effective 2D method to handle 3D data while efficiently embedding the 3D knowledge during training.
Our method generates a super-resolution image by stitching slices side by side in the 3D image.
While attaining equal, if not superior, results to 3D networks utilizing only 2D counterparts, the model complexity is reduced by around threefold.
arXiv Detail & Related papers (2022-05-05T09:59:03Z) - 3D Self-Supervised Methods for Medical Imaging [7.65168530693281]
We propose 3D versions for five different self-supervised methods, in the form of proxy tasks.
Our methods facilitate neural network feature learning from unlabeled 3D images, aiming to reduce the required cost for expert annotation.
The developed algorithms are 3D Contrastive Predictive Coding, 3D Rotation prediction, 3D Jigsaw puzzles, Relative 3D patch location, and 3D Exemplar networks.
arXiv Detail & Related papers (2020-06-06T09:56:58Z) - Exemplar Fine-Tuning for 3D Human Model Fitting Towards In-the-Wild 3D
Human Pose Estimation [107.07047303858664]
Large-scale human datasets with 3D ground-truth annotations are difficult to obtain in the wild.
We address this problem by augmenting existing 2D datasets with high-quality 3D pose fits.
The resulting annotations are sufficient to train from scratch 3D pose regressor networks that outperform the current state-of-the-art on in-the-wild benchmarks.
arXiv Detail & Related papers (2020-04-07T20:21:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.