3D-GMIC: an efficient deep neural network to find small objects in large
3D images
- URL: http://arxiv.org/abs/2210.08645v1
- Date: Sun, 16 Oct 2022 21:58:54 GMT
- Title: 3D-GMIC: an efficient deep neural network to find small objects in large
3D images
- Authors: Jungkyu Park, Jakub Ch{\l}\k{e}dowski, Stanis{\l}aw Jastrz\k{e}bski,
Jan Witowski, Yanqi Xu, Linda Du, Sushma Gaddam, Eric Kim, Alana Lewin, Ujas
Parikh, Anastasia Plaunova, Sardius Chen, Alexandra Millet, James Park,
Kristine Pysarenko, Shalin Patel, Julia Goldberg, Melanie Wegener, Linda Moy,
Laura Heacock, Beatriu Reig, Krzysztof J. Geras
- Abstract summary: 3D imaging enables a more accurate diagnosis by providing spatial information about organ anatomy.
Using 3D images to train AI models is computationally challenging because they consist of tens or hundreds of times more pixels than their 2D counterparts.
We propose a novel neural network architecture that enables computationally efficient classification of 3D medical images in their full resolution.
- Score: 41.334361182700164
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D imaging enables a more accurate diagnosis by providing spatial information
about organ anatomy. However, using 3D images to train AI models is
computationally challenging because they consist of tens or hundreds of times
more pixels than their 2D counterparts. To train with high-resolution 3D
images, convolutional neural networks typically resort to downsampling them or
projecting them to two dimensions. In this work, we propose an effective
alternative, a novel neural network architecture that enables computationally
efficient classification of 3D medical images in their full resolution.
Compared to off-the-shelf convolutional neural networks, 3D-GMIC uses
77.98%-90.05% less GPU memory and 91.23%-96.02% less computation. While our
network is trained only with image-level labels, without segmentation labels,
it explains its classification predictions by providing pixel-level saliency
maps. On a dataset collected at NYU Langone Health, including 85,526 patients
with full-field 2D mammography (FFDM), synthetic 2D mammography, and 3D
mammography (DBT), our model, the 3D Globally-Aware Multiple Instance
Classifier (3D-GMIC), achieves a breast-wise AUC of 0.831 (95% CI: 0.769-0.887)
in classifying breasts with malignant findings using DBT images. As DBT and 2D
mammography capture different information, averaging predictions on 2D and 3D
mammography together leads to a diverse ensemble with an improved breast-wise
AUC of 0.841 (95% CI: 0.768-0.895). Our model generalizes well to an external
dataset from Duke University Hospital, achieving an image-wise AUC of 0.848
(95% CI: 0.798-0.896) in classifying DBT images with malignant findings.
Related papers
- Medical Slice Transformer: Improved Diagnosis and Explainability on 3D Medical Images with DINOv2 [1.6275928583134276]
We introduce the Medical Slice Transformer (MST) framework to adapt 2D self-supervised models for 3D medical image analysis.
MST offers enhanced diagnostic accuracy and explainability compared to convolutional neural networks.
arXiv Detail & Related papers (2024-11-24T12:11:11Z) - Decomposing 3D Neuroimaging into 2+1D Processing for Schizophrenia
Recognition [25.80846093248797]
We propose to process the 3D data by a 2+1D framework so that we can exploit the powerful deep 2D Convolutional Neural Network (CNN) networks pre-trained on the huge ImageNet dataset for 3D neuroimaging recognition.
Specifically, 3D volumes of Magnetic Resonance Imaging (MRI) metrics are decomposed to 2D slices according to neighboring voxel positions.
Global pooling is applied to remove redundant information as the activation patterns are sparsely distributed over feature maps.
Channel-wise and slice-wise convolutions are proposed to aggregate the contextual information in the third dimension unprocessed by the 2D CNN model.
arXiv Detail & Related papers (2022-11-21T15:22:59Z) - Super Images -- A New 2D Perspective on 3D Medical Imaging Analysis [0.0]
We present a simple yet effective 2D method to handle 3D data while efficiently embedding the 3D knowledge during training.
Our method generates a super-resolution image by stitching slices side by side in the 3D image.
While attaining equal, if not superior, results to 3D networks utilizing only 2D counterparts, the model complexity is reduced by around threefold.
arXiv Detail & Related papers (2022-05-05T09:59:03Z) - 3D Convolutional Neural Networks for Stalled Brain Capillary Detection [72.21315180830733]
Brain vasculature dysfunctions such as stalled blood flow in cerebral capillaries are associated with cognitive decline and pathogenesis in Alzheimer's disease.
Here, we describe a deep learning-based approach for automatic detection of stalled capillaries in brain images based on 3D convolutional neural networks.
In this setting, our approach outperformed other methods and demonstrated state-of-the-art results, achieving 0.85 Matthews correlation coefficient, 85% sensitivity, and 99.3% specificity.
arXiv Detail & Related papers (2021-04-04T20:30:14Z) - Automated Model Design and Benchmarking of 3D Deep Learning Models for
COVID-19 Detection with Chest CT Scans [72.04652116817238]
We propose a differentiable neural architecture search (DNAS) framework to automatically search for the 3D DL models for 3D chest CT scans classification.
We also exploit the Class Activation Mapping (CAM) technique on our models to provide the interpretability of the results.
arXiv Detail & Related papers (2021-01-14T03:45:01Z) - Revisiting 3D Context Modeling with Supervised Pre-training for
Universal Lesion Detection in CT Slices [48.85784310158493]
We propose a Modified Pseudo-3D Feature Pyramid Network (MP3D FPN) to efficiently extract 3D context enhanced 2D features for universal lesion detection in CT slices.
With the novel pre-training method, the proposed MP3D FPN achieves state-of-the-art detection performance on the DeepLesion dataset.
The proposed 3D pre-trained weights can potentially be used to boost the performance of other 3D medical image analysis tasks.
arXiv Detail & Related papers (2020-12-16T07:11:16Z) - PerMO: Perceiving More at Once from a Single Image for Autonomous
Driving [76.35684439949094]
We present a novel approach to detect, segment, and reconstruct complete textured 3D models of vehicles from a single image.
Our approach combines the strengths of deep learning and the elegance of traditional techniques.
We have integrated these algorithms with an autonomous driving system.
arXiv Detail & Related papers (2020-07-16T05:02:45Z) - 3D Self-Supervised Methods for Medical Imaging [7.65168530693281]
We propose 3D versions for five different self-supervised methods, in the form of proxy tasks.
Our methods facilitate neural network feature learning from unlabeled 3D images, aiming to reduce the required cost for expert annotation.
The developed algorithms are 3D Contrastive Predictive Coding, 3D Rotation prediction, 3D Jigsaw puzzles, Relative 3D patch location, and 3D Exemplar networks.
arXiv Detail & Related papers (2020-06-06T09:56:58Z) - 2.75D: Boosting learning by representing 3D Medical imaging to 2D
features for small data [54.223614679807994]
3D convolutional neural networks (CNNs) have started to show superior performance to 2D CNNs in numerous deep learning tasks.
Applying transfer learning on 3D CNN is challenging due to a lack of publicly available pre-trained 3D models.
In this work, we proposed a novel 2D strategical representation of volumetric data, namely 2.75D.
As a result, 2D CNN networks can also be used to learn volumetric information.
arXiv Detail & Related papers (2020-02-11T08:24:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.