FCN-Transformer Feature Fusion for Polyp Segmentation
- URL: http://arxiv.org/abs/2208.08352v1
- Date: Wed, 17 Aug 2022 15:31:06 GMT
- Title: FCN-Transformer Feature Fusion for Polyp Segmentation
- Authors: Edward Sanderson and Bogdan J. Matuszewski
- Abstract summary: Colonoscopy is widely recognised as the gold standard procedure for the early detection of colorectal cancer.
The manual segmentation of polyps in colonoscopy images is time-consuming.
The use of deep learning for automation of polyp segmentation has become important.
- Score: 12.62213319797323
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Colonoscopy is widely recognised as the gold standard procedure for the early
detection of colorectal cancer (CRC). Segmentation is valuable for two
significant clinical applications, namely lesion detection and classification,
providing means to improve accuracy and robustness. The manual segmentation of
polyps in colonoscopy images is time-consuming. As a result, the use of deep
learning (DL) for automation of polyp segmentation has become important.
However, DL-based solutions can be vulnerable to overfitting and the resulting
inability to generalise to images captured by different colonoscopes. Recent
transformer-based architectures for semantic segmentation both achieve higher
performance and generalise better than alternatives, however typically predict
a segmentation map of $\frac{h}{4}\times\frac{w}{4}$ spatial dimensions for a
$h\times w$ input image. To this end, we propose a new architecture for
full-size segmentation which leverages the strengths of a transformer in
extracting the most important features for segmentation in a primary branch,
while compensating for its limitations in full-size prediction with a secondary
fully convolutional branch. The resulting features from both branches are then
fused for final prediction of a $h\times w$ segmentation map. We demonstrate
our method's state-of-the-art performance with respect to the mDice, mIoU,
mPrecision, and mRecall metrics, on both the Kvasir-SEG and CVC-ClinicDB
dataset benchmarks. Additionally, we train the model on each of these datasets
and evaluate on the other to demonstrate its superior generalisation
performance.
Related papers
- Prompting Segment Anything Model with Domain-Adaptive Prototype for Generalizable Medical Image Segmentation [49.5901368256326]
We propose a novel Domain-Adaptive Prompt framework for fine-tuning the Segment Anything Model (termed as DAPSAM) in segmenting medical images.
Our DAPSAM achieves state-of-the-art performance on two medical image segmentation tasks with different modalities.
arXiv Detail & Related papers (2024-09-19T07:28:33Z) - ASPS: Augmented Segment Anything Model for Polyp Segmentation [77.25557224490075]
The Segment Anything Model (SAM) has introduced unprecedented potential for polyp segmentation.
SAM's Transformer-based structure prioritizes global and low-frequency information.
CFA integrates a trainable CNN encoder branch with a frozen ViT encoder, enabling the integration of domain-specific knowledge.
arXiv Detail & Related papers (2024-06-30T14:55:32Z) - Adaptation of Distinct Semantics for Uncertain Areas in Polyp Segmentation [11.646574658785362]
This work presents a new novel architecture namely Adaptation of Distinct Semantics for Uncertain Areas in Polyp (ADSNet)
ADSNet modifies misclassified details and recovers weak features having the ability to vanish and not be detected at the final stage.
experimental results demonstrate the great correction and recovery ability leading to better segmentation performance compared to the other state of the art in the polyp image segmentation task.
arXiv Detail & Related papers (2024-05-13T07:41:28Z) - Dual-scale Enhanced and Cross-generative Consistency Learning for Semi-supervised Medical Image Segmentation [49.57907601086494]
Medical image segmentation plays a crucial role in computer-aided diagnosis.
We propose a novel Dual-scale Enhanced and Cross-generative consistency learning framework for semi-supervised medical image (DEC-Seg)
arXiv Detail & Related papers (2023-12-26T12:56:31Z) - Edge-aware Feature Aggregation Network for Polyp Segmentation [40.3881565207086]
In this study, we present a novel Edge-aware Feature Aggregation Network (EFA-Net) for polyp segmentation.
EFA-Net can fully make use of cross-level and multi-scale features to enhance the performance of polyp segmentation.
Experimental results on five widely adopted colonoscopy datasets show that our EFA-Net outperforms state-of-the-art polyp segmentation methods in terms of generalization and effectiveness.
arXiv Detail & Related papers (2023-09-19T11:09:38Z) - Interactive Segmentation as Gaussian Process Classification [58.44673380545409]
Click-based interactive segmentation (IS) aims to extract the target objects under user interaction.
Most of the current deep learning (DL)-based methods mainly follow the general pipelines of semantic segmentation.
We propose to formulate the IS task as a Gaussian process (GP)-based pixel-wise binary classification model on each image.
arXiv Detail & Related papers (2023-02-28T14:01:01Z) - FCB-SwinV2 Transformer for Polyp Segmentation [0.0]
Polyp segmentation within colonoscopy video frames using deep learning models has the potential to automate the workflow of clinicians.
Recent state-of-the-art deep learning polyp segmentation models have combined the outputs of Fully Convolutional Network architectures and Transformer Network architectures which work in parallel.
In this paper we propose modifications to the current state-of-the-art polyp segmentation model FCBFormer.
The performance of the FCB-SwinV2 Transformer is evaluated on the popular colonoscopy segmentation bench-marking datasets.
arXiv Detail & Related papers (2023-02-02T11:42:26Z) - Deep ensembles based on Stochastic Activation Selection for Polyp
Segmentation [82.61182037130406]
This work deals with medical image segmentation and in particular with accurate polyp detection and segmentation during colonoscopy examinations.
Basic architecture in image segmentation consists of an encoder and a decoder.
We compare some variant of the DeepLab architecture obtained by varying the decoder backbone.
arXiv Detail & Related papers (2021-04-02T02:07:37Z) - PraNet: Parallel Reverse Attention Network for Polyp Segmentation [155.93344756264824]
We propose a parallel reverse attention network (PraNet) for accurate polyp segmentation in colonoscopy images.
We first aggregate the features in high-level layers using a parallel partial decoder (PPD)
In addition, we mine the boundary cues using a reverse attention (RA) module, which is able to establish the relationship between areas and boundary cues.
arXiv Detail & Related papers (2020-06-13T08:13:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.