Revisiting 3D Context Modeling with Supervised Pre-training for
Universal Lesion Detection in CT Slices
- URL: http://arxiv.org/abs/2012.08770v1
- Date: Wed, 16 Dec 2020 07:11:16 GMT
- Title: Revisiting 3D Context Modeling with Supervised Pre-training for
Universal Lesion Detection in CT Slices
- Authors: Shu Zhang, Jincheng Xu, Yu-Chun Chen, Jiechao Ma, Zihao Li, Yizhou
Wang and Yizhou Yu
- Abstract summary: We propose a Modified Pseudo-3D Feature Pyramid Network (MP3D FPN) to efficiently extract 3D context enhanced 2D features for universal lesion detection in CT slices.
With the novel pre-training method, the proposed MP3D FPN achieves state-of-the-art detection performance on the DeepLesion dataset.
The proposed 3D pre-trained weights can potentially be used to boost the performance of other 3D medical image analysis tasks.
- Score: 48.85784310158493
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Universal lesion detection from computed tomography (CT) slices is important
for comprehensive disease screening. Since each lesion can locate in multiple
adjacent slices, 3D context modeling is of great significance for developing
automated lesion detection algorithms. In this work, we propose a Modified
Pseudo-3D Feature Pyramid Network (MP3D FPN) that leverages depthwise separable
convolutional filters and a group transform module (GTM) to efficiently extract
3D context enhanced 2D features for universal lesion detection in CT slices. To
facilitate faster convergence, a novel 3D network pre-training method is
derived using solely large-scale 2D object detection dataset in the natural
image domain. We demonstrate that with the novel pre-training method, the
proposed MP3D FPN achieves state-of-the-art detection performance on the
DeepLesion dataset (3.48% absolute improvement in the sensitivity of FPs@0.5),
significantly surpassing the baseline method by up to 6.06% (in MAP@0.5) which
adopts 2D convolution for 3D context modeling. Moreover, the proposed 3D
pre-trained weights can potentially be used to boost the performance of other
3D medical image analysis tasks.
Related papers
- Persistence Image from 3D Medical Image: Superpixel and Optimized Gaussian Coefficient [3.808587330262038]
Topological data analysis (TDA) uncovers crucial properties of objects in medical imaging.
Previous research primarily focused on 2D image analysis, neglecting the comprehensive 3D context.
We propose an innovative 3D TDA approach that incorporates the concept of superpixels to transform 3D medical image features into point cloud data.
arXiv Detail & Related papers (2024-08-15T03:24:00Z) - Diff3Dformer: Leveraging Slice Sequence Diffusion for Enhanced 3D CT Classification with Transformer Networks [5.806035963947936]
We propose a Diffusion-based 3D Vision Transformer (Diff3Dformer) to aggregate repetitive information within 3D CT scans.
Our method exhibits improved performance on two different scales of small datasets of 3D lung CT scans.
arXiv Detail & Related papers (2024-06-24T23:23:18Z) - Geometric Transformation Uncertainty for Improving 3D Fetal Brain Pose Prediction from Freehand 2D Ultrasound Videos [0.8579241568505183]
We propose an uncertainty-aware deep learning model for automated 3D plane localization in 2D fetal brain images.
Our proposed method, QAERTS, demonstrates superior pose estimation accuracy than the state-of-the-art and most of the uncertainty-based approaches.
arXiv Detail & Related papers (2024-05-21T22:42:08Z) - 3DiffTection: 3D Object Detection with Geometry-Aware Diffusion Features [70.50665869806188]
3DiffTection is a state-of-the-art method for 3D object detection from single images.
We fine-tune a diffusion model to perform novel view synthesis conditioned on a single image.
We further train the model on target data with detection supervision.
arXiv Detail & Related papers (2023-11-07T23:46:41Z) - Complementary Pseudo Multimodal Feature for Point Cloud Anomaly
Detection [5.059800023492045]
Point cloud (PCD) anomaly detection steadily emerges as a promising research area.
This study proposes Complementary Pseudo Multimodal Feature (CPMF) that incorporates local geometrical information in 3D modality.
Experiments demonstrate the complementary capacity between 2D and 3D modality features, with 95.15% image-level AU-ROC and 92.93% pixel-level PRO.
arXiv Detail & Related papers (2023-03-23T11:52:17Z) - Improving 3D Imaging with Pre-Trained Perpendicular 2D Diffusion Models [52.529394863331326]
We propose a novel approach using two perpendicular pre-trained 2D diffusion models to solve the 3D inverse problem.
Our method is highly effective for 3D medical image reconstruction tasks, including MRI Z-axis super-resolution, compressed sensing MRI, and sparse-view CT.
arXiv Detail & Related papers (2023-03-15T08:28:06Z) - View-Disentangled Transformer for Brain Lesion Detection [50.4918615815066]
We propose a novel view-disentangled transformer to enhance the extraction of MRI features for more accurate tumour detection.
First, the proposed transformer harvests long-range correlation among different positions in a 3D brain scan.
Second, the transformer models a stack of slice features as multiple 2D views and enhance these features view-by-view.
Third, we deploy the proposed transformer module in a transformer backbone, which can effectively detect the 2D regions surrounding brain lesions.
arXiv Detail & Related papers (2022-09-20T11:58:23Z) - Simultaneous Alignment and Surface Regression Using Hybrid 2D-3D
Networks for 3D Coherent Layer Segmentation of Retina OCT Images [33.99874168018807]
In this study, a novel framework based on hybrid 2D-3D convolutional neural networks (CNNs) is proposed to obtain continuous 3D retinal layer surfaces from OCT.
Our framework achieves superior results to state-of-the-art 2D methods in terms of both layer segmentation accuracy and cross-B-scan 3D continuity.
arXiv Detail & Related papers (2022-03-04T15:55:09Z) - Automated Model Design and Benchmarking of 3D Deep Learning Models for
COVID-19 Detection with Chest CT Scans [72.04652116817238]
We propose a differentiable neural architecture search (DNAS) framework to automatically search for the 3D DL models for 3D chest CT scans classification.
We also exploit the Class Activation Mapping (CAM) technique on our models to provide the interpretability of the results.
arXiv Detail & Related papers (2021-01-14T03:45:01Z) - Reinforced Axial Refinement Network for Monocular 3D Object Detection [160.34246529816085]
Monocular 3D object detection aims to extract the 3D position and properties of objects from a 2D input image.
Conventional approaches sample 3D bounding boxes from the space and infer the relationship between the target object and each of them, however, the probability of effective samples is relatively small in the 3D space.
We propose to start with an initial prediction and refine it gradually towards the ground truth, with only one 3d parameter changed in each step.
This requires designing a policy which gets a reward after several steps, and thus we adopt reinforcement learning to optimize it.
arXiv Detail & Related papers (2020-08-31T17:10:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.