3D unsupervised anomaly detection and localization through virtual
multi-view projection and reconstruction: Clinical validation on low-dose
chest computed tomography
- URL: http://arxiv.org/abs/2206.13385v1
- Date: Sat, 18 Jun 2022 13:22:00 GMT
- Title: 3D unsupervised anomaly detection and localization through virtual
multi-view projection and reconstruction: Clinical validation on low-dose
chest computed tomography
- Authors: Kyung-Su Kim, Seong Je Oh, Ju Hwan Lee, Myung Jin Chung
- Abstract summary: We propose a method based on a deep neural network for computer-aided diagnosis called virtual multi-view projection and reconstruction.
The proposed method improves the patient-level anomaly detection by 10% compared with a gold standard based on supervised learning.
It localizes the anomaly region with 93% accuracy, demonstrating its high performance.
- Score: 2.2302915692528367
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Computer-aided diagnosis for low-dose computed tomography (CT) based on deep
learning has recently attracted attention as a first-line automatic testing
tool because of its high accuracy and low radiation exposure. However, existing
methods rely on supervised learning, imposing an additional burden to doctors
for collecting disease data or annotating spatial labels for network training,
consequently hindering their implementation. We propose a method based on a
deep neural network for computer-aided diagnosis called virtual multi-view
projection and reconstruction for unsupervised anomaly detection. Presumably,
this is the first method that only requires data from healthy patients for
training to identify three-dimensional (3D) regions containing any anomalies.
The method has three key components. Unlike existing computer-aided diagnosis
tools that use conventional CT slices as the network input, our method 1)
improves the recognition of 3D lung structures by virtually projecting an
extracted 3D lung region to obtain two-dimensional (2D) images from diverse
views to serve as network inputs, 2) accommodates the input diversity gain for
accurate anomaly detection, and 3) achieves 3D anomaly/disease localization
through a novel 3D map restoration method using multiple 2D anomaly maps. The
proposed method based on unsupervised learning improves the patient-level
anomaly detection by 10% (area under the curve, 0.959) compared with a gold
standard based on supervised learning (area under the curve, 0.848), and it
localizes the anomaly region with 93% accuracy, demonstrating its high
performance.
Related papers
- Abnormality-Driven Representation Learning for Radiology Imaging [0.8321462983924758]
We introduce lesion-enhanced contrastive learning (LeCL), a novel approach to obtain visual representations driven by abnormalities in 2D axial slices across different locations of the CT scans.
We evaluate our approach across three clinical tasks: tumor lesion location, lung disease detection, and patient staging, benchmarking against four state-of-the-art foundation models.
arXiv Detail & Related papers (2024-11-25T13:53:26Z) - Adapting Visual-Language Models for Generalizable Anomaly Detection in Medical Images [68.42215385041114]
This paper introduces a novel lightweight multi-level adaptation and comparison framework to repurpose the CLIP model for medical anomaly detection.
Our approach integrates multiple residual adapters into the pre-trained visual encoder, enabling a stepwise enhancement of visual features across different levels.
Our experiments on medical anomaly detection benchmarks demonstrate that our method significantly surpasses current state-of-the-art models.
arXiv Detail & Related papers (2024-03-19T09:28:19Z) - Self-supervised Feature Adaptation for 3D Industrial Anomaly Detection [59.41026558455904]
We focus on multi-modal anomaly detection. Specifically, we investigate early multi-modal approaches that attempted to utilize models pre-trained on large-scale visual datasets.
We propose a Local-to-global Self-supervised Feature Adaptation (LSFA) method to finetune the adaptors and learn task-oriented representation toward anomaly detection.
arXiv Detail & Related papers (2024-01-06T07:30:41Z) - Multi-View Vertebra Localization and Identification from CT Images [57.56509107412658]
We propose a multi-view vertebra localization and identification from CT images.
We convert the 3D problem into a 2D localization and identification task on different views.
Our method can learn the multi-view global information naturally.
arXiv Detail & Related papers (2023-07-24T14:43:07Z) - Explainable multiple abnormality classification of chest CT volumes with
AxialNet and HiResCAM [89.2175350956813]
We introduce the challenging new task of explainable multiple abnormality classification in volumetric medical images.
We propose a multiple instance learning convolutional neural network, AxialNet, that allows identification of top slices for each abnormality.
We then aim to improve the model's learning through a novel mask loss that leverages HiResCAM and 3D allowed regions.
arXiv Detail & Related papers (2021-11-24T01:14:33Z) - Delving into Localization Errors for Monocular 3D Object Detection [85.77319416168362]
Estimating 3D bounding boxes from monocular images is an essential component in autonomous driving.
In this work, we quantify the impact introduced by each sub-task and find the localization error' is the vital factor in restricting monocular 3D detection.
arXiv Detail & Related papers (2021-03-30T10:38:01Z) - Automated 3D cephalometric landmark identification using computerized
tomography [1.4349468613117398]
Identification of 3D cephalometric landmarks that serve as proxy to the shape of human skull is the fundamental step in cephalometric analysis.
Recently, automatic landmarking of 2D cephalograms using deep learning (DL) has achieved great success, but 3D landmarking for more than 80 landmarks has not yet reached a satisfactory level.
This paper presents a semi-supervised DL method for 3D landmarking that takes advantage of anonymized landmark dataset with paired CT data being removed.
arXiv Detail & Related papers (2020-12-16T07:29:32Z) - Revisiting 3D Context Modeling with Supervised Pre-training for
Universal Lesion Detection in CT Slices [48.85784310158493]
We propose a Modified Pseudo-3D Feature Pyramid Network (MP3D FPN) to efficiently extract 3D context enhanced 2D features for universal lesion detection in CT slices.
With the novel pre-training method, the proposed MP3D FPN achieves state-of-the-art detection performance on the DeepLesion dataset.
The proposed 3D pre-trained weights can potentially be used to boost the performance of other 3D medical image analysis tasks.
arXiv Detail & Related papers (2020-12-16T07:11:16Z) - Deep Volumetric Universal Lesion Detection using Light-Weight Pseudo 3D
Convolution and Surface Point Regression [23.916776570010285]
Computer-aided lesion/significant-findings detection techniques are at the core of medical imaging.
We propose a novel deep anchor-free one-stage VULD framework that incorporates (1) P3DC operators to recycle the architectural configurations and pre-trained weights from the off-the-shelf 2D networks.
New SPR method to effectively regress the 3D lesion spatial extents by pinpointing their representative key points on lesion surfaces.
arXiv Detail & Related papers (2020-08-30T19:42:06Z) - Keypoints Localization for Joint Vertebra Detection and Fracture
Severity Quantification [0.04925906256430176]
Vertebral body compression fractures are reliable early signs of osteoporosis.
We propose a new two-step algorithm to localize the vertebral column in 3D CT images.
We simultaneously detect individual vertebrae and quantify fractures in 2D.
arXiv Detail & Related papers (2020-05-25T08:05:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.