3D Axial-Attention for Lung Nodule Classification
- URL: http://arxiv.org/abs/2012.14117v2
- Date: Sat, 2 Jan 2021 06:52:30 GMT
- Title: 3D Axial-Attention for Lung Nodule Classification
- Authors: Mundher Al-Shabi, Kelvin Shak, Maxine Tan
- Abstract summary: We propose to use 3D Axial-Attention, which requires a fraction of the computing power of a regular Non-Local network.
We solve the position invariant problem of the Non-Local network by proposing adding 3D positional encoding to shared embeddings.
Our results show that the 3D Axial-Attention model achieves state-of-the-art performance on all evaluation metrics including AUC and Accuracy.
- Score: 0.11458853556386794
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Purpose: In recent years, Non-Local based methods have been successfully
applied to lung nodule classification. However, these methods offer 2D
attention or a limited 3D attention to low-resolution feature maps. Moreover,
they still depend on a convenient local filter such as convolution as full 3D
attention is expensive to compute and requires a big dataset, which might not
be available. Methods: We propose to use 3D Axial-Attention, which requires a
fraction of the computing power of a regular Non-Local network. Additionally,
we solve the position invariant problem of the Non-Local network by proposing
adding 3D positional encoding to shared embeddings. Results: We validated the
proposed method on the LIDC-IDRI dataset by following a rigorous experimental
setup using only nodules annotated by at least three radiologists. Our results
show that the 3D Axial-Attention model achieves state-of-the-art performance on
all evaluation metrics including AUC and Accuracy. Conclusions: The proposed
model provides full 3D attention effectively, which can be used in all layers
without the need for local filters. The experimental results show the
importance of full 3D attention for classifying lung nodules.
Related papers
- CatFree3D: Category-agnostic 3D Object Detection with Diffusion [63.75470913278591]
We introduce a novel pipeline that decouples 3D detection from 2D detection and depth prediction.
We also introduce the Normalised Hungarian Distance (NHD) metric for an accurate evaluation of 3D detection results.
arXiv Detail & Related papers (2024-08-22T22:05:57Z) - AttentNet: Fully Convolutional 3D Attention for Lung Nodule Detection [4.47196217712431]
In 3D medical imaging, such as 3D pulmonary CT scans, efficient processing is crucial due to the large data volume.
We investigate the possibility of incorporating fully convolutional (conv.) attention in 3D context.
We present two 3D fully conv. attention blocks, demonstrating their effectiveness in 3D context.
arXiv Detail & Related papers (2024-07-19T17:06:27Z) - Unsupervised 3D out-of-distribution detection with latent diffusion
models [1.7587591581995812]
We propose to use Latent Diffusion Models (LDMs) to enable the scaling of DDPMs to high-resolution 3D medical data.
Not only does the proposed LDM-based approach achieve statistically significant better performance, it also shows less sensitivity to the underlying latent representation.
arXiv Detail & Related papers (2023-07-07T18:00:38Z) - Probabilistic 3D segmentation for aleatoric uncertainty quantification
in full 3D medical data [7.615431940103322]
We develop a 3D probabilistic segmentation framework augmented with Normalizing Flows.
We are the first to present a 3D Squared Generalized Energy Distance (GED) of 0.401 and a high 0.468 Hungarian-matched 3D IoU.
arXiv Detail & Related papers (2023-05-01T17:19:20Z) - Homography Loss for Monocular 3D Object Detection [54.04870007473932]
A differentiable loss function, termed as Homography Loss, is proposed to achieve the goal, which exploits both 2D and 3D information.
Our method yields the best performance compared with the other state-of-the-arts by a large margin on KITTI 3D datasets.
arXiv Detail & Related papers (2022-04-02T03:48:03Z) - Shape Prior Non-Uniform Sampling Guided Real-time Stereo 3D Object
Detection [59.765645791588454]
Recently introduced RTS3D builds an efficient 4D Feature-Consistency Embedding space for the intermediate representation of object without depth supervision.
We propose a shape prior non-uniform sampling strategy that performs dense sampling in outer region and sparse sampling in inner region.
Our proposed method has 2.57% improvement on AP3d almost without extra network parameters.
arXiv Detail & Related papers (2021-06-18T09:14:55Z) - R2U3D: Recurrent Residual 3D U-Net for Lung Segmentation [17.343802171952195]
We propose a novel model, namely, Recurrent Residual 3D U-Net (R2U3D), for the 3D lung segmentation task.
In particular, the proposed model integrates 3D convolution into the Recurrent Residual Neural Network based on U-Net.
The proposed R2U3D network is trained on the publicly available dataset LUNA16 and it achieves state-of-the-art performance.
arXiv Detail & Related papers (2021-05-05T19:17:14Z) - SCPM-Net: An Anchor-free 3D Lung Nodule Detection Network using Sphere
Representation and Center Points Matching [47.79483848496141]
We propose a 3D sphere representation-based center-points matching detection network (SCPM-Net)
It is anchor-free and automatically predicts the position, radius, and offset of nodules without the manual design of nodule/anchor parameters.
We show that our proposed SCPM-Net framework achieves superior performance compared with existing used anchor-based and anchor-free methods for lung nodule detection.
arXiv Detail & Related papers (2021-04-12T05:51:29Z) - Delving into Localization Errors for Monocular 3D Object Detection [85.77319416168362]
Estimating 3D bounding boxes from monocular images is an essential component in autonomous driving.
In this work, we quantify the impact introduced by each sub-task and find the localization error' is the vital factor in restricting monocular 3D detection.
arXiv Detail & Related papers (2021-03-30T10:38:01Z) - Reinforced Axial Refinement Network for Monocular 3D Object Detection [160.34246529816085]
Monocular 3D object detection aims to extract the 3D position and properties of objects from a 2D input image.
Conventional approaches sample 3D bounding boxes from the space and infer the relationship between the target object and each of them, however, the probability of effective samples is relatively small in the 3D space.
We propose to start with an initial prediction and refine it gradually towards the ground truth, with only one 3d parameter changed in each step.
This requires designing a policy which gets a reward after several steps, and thus we adopt reinforcement learning to optimize it.
arXiv Detail & Related papers (2020-08-31T17:10:48Z) - D3Feat: Joint Learning of Dense Detection and Description of 3D Local
Features [51.04841465193678]
We leverage a 3D fully convolutional network for 3D point clouds.
We propose a novel and practical learning mechanism that densely predicts both a detection score and a description feature for each 3D point.
Our method achieves state-of-the-art results in both indoor and outdoor scenarios.
arXiv Detail & Related papers (2020-03-06T12:51:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.