Deep Learning-based Biological Anatomical Landmark Detection in
Colonoscopy Videos
- URL: http://arxiv.org/abs/2108.02948v1
- Date: Fri, 6 Aug 2021 05:52:32 GMT
- Title: Deep Learning-based Biological Anatomical Landmark Detection in
Colonoscopy Videos
- Authors: Kaiwei Che, Chengwei Ye, Yibing Yao, Nachuan Ma, Ruo Zhang, Jiankun
Wang, and Max Q.-H. Meng
- Abstract summary: We propose a novel deep learning-based approach to detect biological anatomical landmarks in colonoscopy videos.
Average detection accuracy reaches 99.75%, while the average IoU of 0.91 shows a high degree of similarity between our predicted landmark periods and ground truth.
- Score: 21.384094148149003
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Colonoscopy is a standard imaging tool for visualizing the entire
gastrointestinal (GI) tract of patients to capture lesion areas. However, it
takes the clinicians excessive time to review a large number of images
extracted from colonoscopy videos. Thus, automatic detection of biological
anatomical landmarks within the colon is highly demanded, which can help reduce
the burden of clinicians by providing guidance information for the locations of
lesion areas. In this article, we propose a novel deep learning-based approach
to detect biological anatomical landmarks in colonoscopy videos. First, raw
colonoscopy video sequences are pre-processed to reject interference frames.
Second, a ResNet-101 based network is used to detect three biological
anatomical landmarks separately to obtain the intermediate detection results.
Third, to achieve more reliable localization of the landmark periods within the
whole video period, we propose to post-process the intermediate detection
results by identifying the incorrectly predicted frames based on their temporal
distribution and reassigning them back to the correct class. Finally, the
average detection accuracy reaches 99.75\%. Meanwhile, the average IoU of 0.91
shows a high degree of similarity between our predicted landmark periods and
ground truth. The experimental results demonstrate that our proposed model is
capable of accurately detecting and localizing biological anatomical landmarks
from colonoscopy videos.
Related papers
- A novel open-source ultrasound dataset with deep learning benchmarks for
spinal cord injury localization and anatomical segmentation [1.02101998415327]
We present an ultrasound dataset of 10,223-mode (B-mode) images consisting of sagittal slices of porcine spinal cords.
We benchmark the performance metrics of several state-of-the-art object detection algorithms to localize the site of injury.
We evaluate the zero-shot generalization capabilities of the segmentation models on human ultrasound spinal cord images.
arXiv Detail & Related papers (2024-09-24T20:22:59Z) - ToDER: Towards Colonoscopy Depth Estimation and Reconstruction with Geometry Constraint Adaptation [67.22294293695255]
We propose a novel reconstruction pipeline with a bi-directional adaptation architecture named ToDER to get precise depth estimations.
Experimental results demonstrate that our approach can precisely predict depth maps in both realistic and synthetic colonoscopy videos.
arXiv Detail & Related papers (2024-07-23T14:24:26Z) - Spatial-aware Attention Generative Adversarial Network for Semi-supervised Anomaly Detection in Medical Image [63.59114880750643]
We introduce a novel Spatial-aware Attention Generative Adrialversa Network (SAGAN) for one-class semi-supervised generation of health images.
SAGAN generates high-quality health images corresponding to unlabeled data, guided by the reconstruction of normal images and restoration of pseudo-anomaly images.
Extensive experiments on three medical datasets demonstrate that the proposed SAGAN outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2024-05-21T15:41:34Z) - Real-time guidewire tracking and segmentation in intraoperative x-ray [52.51797358201872]
We propose a two-stage deep learning framework for real-time guidewire segmentation and tracking.
In the first stage, a Yolov5 detector is trained, using the original X-ray images as well as synthetic ones, to output the bounding boxes of possible target guidewires.
In the second stage, a novel and efficient network is proposed to segment the guidewire in each detected bounding box.
arXiv Detail & Related papers (2024-04-12T20:39:19Z) - SoftEnNet: Symbiotic Monocular Depth Estimation and Lumen Segmentation
for Colonoscopy Endorobots [2.9696400288366127]
Colorectal cancer is the third most common cause of cancer death worldwide.
A vision-based autonomous endorobot can improve colonoscopy procedures significantly.
arXiv Detail & Related papers (2023-01-19T16:22:17Z) - Segmentation of kidney stones in endoscopic video feeds [2.572404739180802]
We describe how we built a dataset from the raw videos and how we developed a pipeline to automate as much of the process as possible.
To show clinical potential for real-time use, we also confirmed that our best trained model can accurately annotate new videos at 30 frames per second.
arXiv Detail & Related papers (2022-04-29T16:00:52Z) - A Deep Learning Approach to Predicting Collateral Flow in Stroke
Patients Using Radiomic Features from Perfusion Images [58.17507437526425]
Collateral circulation results from specialized anastomotic channels which provide oxygenated blood to regions with compromised blood flow.
The actual grading is mostly done through manual inspection of the acquired images.
We present a deep learning approach to predicting collateral flow grading in stroke patients based on radiomic features extracted from MR perfusion data.
arXiv Detail & Related papers (2021-10-24T18:58:40Z) - Siamese Network Features for Endoscopy Image and Video Localization [0.0]
Localizing frames provide valuable information about anomaly location.
In this study, we present a combination of meta-learning and deep learning for localizing both endoscopy images and video.
arXiv Detail & Related papers (2021-03-15T16:24:30Z) - Colonoscopy Polyp Detection: Domain Adaptation From Medical Report
Images to Real-time Videos [76.37907640271806]
We propose an Image-video-joint polyp detection network (Ivy-Net) to address the domain gap between colonoscopy images from historical medical reports and real-time videos.
Experiments on the collected dataset demonstrate that our Ivy-Net achieves the state-of-the-art result on colonoscopy video.
arXiv Detail & Related papers (2020-12-31T10:33:09Z) - Motion-based Camera Localization System in Colonoscopy Videos [7.800211144015489]
We propose a camera localization system to estimate the relative location of the camera and classify the colon into anatomical segments.
The experimental results show that the performance of the proposed method is superior to other published methods.
arXiv Detail & Related papers (2020-12-03T03:57:12Z) - Assisted Probe Positioning for Ultrasound Guided Radiotherapy Using
Image Sequence Classification [55.96221340756895]
Effective transperineal ultrasound image guidance in prostate external beam radiotherapy requires consistent alignment between probe and prostate at each session during patient set-up.
We demonstrate a method for ensuring accurate probe placement through joint classification of images and probe position data.
Using a multi-input multi-task algorithm, spatial coordinate data from an optically tracked ultrasound probe is combined with an image clas-sifier using a recurrent neural network to generate two sets of predictions in real-time.
The algorithm identified optimal probe alignment within a mean (standard deviation) range of 3.7$circ$ (1.2$circ$) from
arXiv Detail & Related papers (2020-10-06T13:55:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.