Accurate Real-time Polyp Detection in Videos from Concatenation of
Latent Features Extracted from Consecutive Frames
- URL: http://arxiv.org/abs/2303.05871v1
- Date: Fri, 10 Mar 2023 11:51:22 GMT
- Title: Accurate Real-time Polyp Detection in Videos from Concatenation of
Latent Features Extracted from Consecutive Frames
- Authors: Hemin Ali Qadir, Younghak Shin, Jacob Bergsland, Ilangko Balasingham
- Abstract summary: Convolutional neural networks (CNNs) are vulnerable to small changes in the input image.
A CNN-based model may miss the same polyp appearing in a series of consecutive frames.
We propose an efficient feature concatenation method for a CNN-based encoder-decoder model.
- Score: 5.2009074009536524
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: An efficient deep learning model that can be implemented in real-time for
polyp detection is crucial to reducing polyp miss-rate during screening
procedures. Convolutional neural networks (CNNs) are vulnerable to small
changes in the input image. A CNN-based model may miss the same polyp appearing
in a series of consecutive frames and produce unsubtle detection output due to
changes in camera pose, lighting condition, light reflection, etc. In this
study, we attempt to tackle this problem by integrating temporal information
among neighboring frames. We propose an efficient feature concatenation method
for a CNN-based encoder-decoder model without adding complexity to the model.
The proposed method incorporates extracted feature maps of previous frames to
detect polyps in the current frame. The experimental results demonstrate that
the proposed method of feature concatenation improves the overall performance
of automatic polyp detection in videos. The following results are obtained on a
public video dataset: sensitivity 90.94\%, precision 90.53\%, and specificity
92.46%
Related papers
- SSTFB: Leveraging self-supervised pretext learning and temporal self-attention with feature branching for real-time video polyp segmentation [4.027361638728112]
We propose a video polyp segmentation method that performs self-supervised learning as an auxiliary task and a spatial-temporal self-attention mechanism for improved representation learning.
Our experimental results demonstrate an improvement with respect to several state-of-the-art (SOTA) methods.
Our ablation study confirms that the choice of the proposed joint end-to-end training improves network accuracy by over 3% and nearly 10% on both the Dice similarity coefficient and intersection-over-union.
arXiv Detail & Related papers (2024-06-14T17:33:11Z) - ECC-PolypDet: Enhanced CenterNet with Contrastive Learning for Automatic
Polyp Detection [88.4359020192429]
Existing methods either involve computationally expensive context aggregation or lack prior modeling of polyps, resulting in poor performance in challenging cases.
In this paper, we propose the Enhanced CenterNet with Contrastive Learning (ECC-PolypDet), a two-stage training & end-to-end inference framework.
Box-assisted Contrastive Learning (BCL) during training to minimize the intra-class difference and maximize the inter-class difference between foreground polyps and backgrounds, enabling our model to capture concealed polyps.
In the fine-tuning stage, we introduce the IoU-guided Sample Re-weighting
arXiv Detail & Related papers (2024-01-10T07:03:41Z) - A Spatial-Temporal Deformable Attention based Framework for Breast
Lesion Detection in Videos [107.96514633713034]
We propose a spatial-temporal deformable attention based framework, named STNet.
Our STNet introduces a spatial-temporal deformable attention module to perform local spatial-temporal feature fusion.
Experiments on the public breast lesion ultrasound video dataset show that our STNet obtains a state-of-the-art detection performance.
arXiv Detail & Related papers (2023-09-09T07:00:10Z) - YONA: You Only Need One Adjacent Reference-frame for Accurate and Fast
Video Polyp Detection [80.68520401539979]
textbfYONA (textbfYou textbfOnly textbfNeed one textbfAdjacent Reference-frame) is an efficient end-to-end training framework for video polyp detection.
Our proposed YONA outperforms previous state-of-the-art competitors by a large margin in both accuracy and speed.
arXiv Detail & Related papers (2023-06-06T13:53:15Z) - Lesion-aware Dynamic Kernel for Polyp Segmentation [49.63274623103663]
We propose a lesion-aware dynamic network (LDNet) for polyp segmentation.
It is a traditional u-shape encoder-decoder structure incorporated with a dynamic kernel generation and updating scheme.
This simple but effective scheme endows our model with powerful segmentation performance and generalization capability.
arXiv Detail & Related papers (2023-01-12T09:53:57Z) - Contrastive Transformer-based Multiple Instance Learning for Weakly
Supervised Polyp Frame Detection [30.51410140271929]
Current polyp detection methods from colonoscopy videos use exclusively normal (i.e., healthy) training images.
We formulate polyp detection as a weakly-supervised anomaly detection task that uses video-level labelled training data to detect frame-level polyps.
arXiv Detail & Related papers (2022-03-23T01:30:48Z) - Exploration of Various Deep Learning Models for Increased Accuracy in
Automatic Polyp Detection [62.997667081978825]
This paper explores deep learning models and algorithms that results in highest accuracy in detecting polyp on colonoscopy images.
Previous studies implemented deep learning using convolution neural network (CNN)
arXiv Detail & Related papers (2022-03-04T04:03:41Z) - Colonoscopy polyp detection with massive endoscopic images [4.458670612147842]
We improved an existing end-to-end polyp detection model with better average precision validated by different data sets.
Our model can achieve state-of-the-art polyp detection performance while still maintain real-time detection speed.
arXiv Detail & Related papers (2022-02-17T16:07:59Z) - Real-time automatic polyp detection in colonoscopy using feature
enhancement module and spatiotemporal similarity correlation unit [34.28382404976628]
State-of-the-art methods are based on convolutional neural networks (CNNs)
Our method combines the two-dimensional (2-D) CNN-based real-time object detector network withtemporal information.
It's demonstrated that our method provides a performance improvement in sensitivity, precision and specificity, and has great potential to be applied in clinical colonoscopy.
arXiv Detail & Related papers (2022-01-25T03:40:30Z) - Colonoscopy Polyp Detection: Domain Adaptation From Medical Report
Images to Real-time Videos [76.37907640271806]
We propose an Image-video-joint polyp detection network (Ivy-Net) to address the domain gap between colonoscopy images from historical medical reports and real-time videos.
Experiments on the collected dataset demonstrate that our Ivy-Net achieves the state-of-the-art result on colonoscopy video.
arXiv Detail & Related papers (2020-12-31T10:33:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.