Polyp-E: Benchmarking the Robustness of Deep Segmentation Models via Polyp Editing
- URL: http://arxiv.org/abs/2410.16732v1
- Date: Tue, 22 Oct 2024 06:30:37 GMT
- Title: Polyp-E: Benchmarking the Robustness of Deep Segmentation Models via Polyp Editing
- Authors: Runpu Wei, Zijin Yin, Kongming Liang, Min Min, Chengwei Pan, Gang Yu, Haonan Huang, Yan Liu, Zhanyu Ma,
- Abstract summary: In daily clinical practice, clinicians exhibit robustness in identifying polyps with both location and size variations.
It is uncertain if deep segmentation models can achieve comparable robustness in automated colonoscopic analysis.
We focus on evaluating the robustness of segmentation models on the polyps with various attributes and healthy samples.
- Score: 32.30835026874521
- License:
- Abstract: Automatic polyp segmentation is helpful to assist clinical diagnosis and treatment. In daily clinical practice, clinicians exhibit robustness in identifying polyps with both location and size variations. It is uncertain if deep segmentation models can achieve comparable robustness in automated colonoscopic analysis. To benchmark the model robustness, we focus on evaluating the robustness of segmentation models on the polyps with various attributes (e.g. location and size) and healthy samples. Based on the Latent Diffusion Model, we perform attribute editing on real polyps and build a new dataset named Polyp-E. Our synthetic dataset boasts exceptional realism, to the extent that clinical experts find it challenging to discern them from real data. We evaluate several existing polyp segmentation models on the proposed benchmark. The results reveal most of the models are highly sensitive to attribute variations. As a novel data augmentation technique, the proposed editing pipeline can improve both in-distribution and out-of-distribution generalization ability. The code and datasets will be released.
Related papers
- Polyp-SES: Automatic Polyp Segmentation with Self-Enriched Semantic Model [7.424888086388194]
We propose an innovative method named Automatic Polyp with Self-Enriched Semantic Model'' to address limitations in existing approaches.
First, we extract a sequence of features from an input image and decode high-level features to generate an initial segmentation mask.
Using the proposed self-enriched semantic module, we query potential semantics and augment deep features with additional semantics, thereby aiding the model in understanding context more effectively.
arXiv Detail & Related papers (2024-10-02T03:34:23Z) - ASPS: Augmented Segment Anything Model for Polyp Segmentation [77.25557224490075]
The Segment Anything Model (SAM) has introduced unprecedented potential for polyp segmentation.
SAM's Transformer-based structure prioritizes global and low-frequency information.
CFA integrates a trainable CNN encoder branch with a frozen ViT encoder, enabling the integration of domain-specific knowledge.
arXiv Detail & Related papers (2024-06-30T14:55:32Z) - ECC-PolypDet: Enhanced CenterNet with Contrastive Learning for Automatic
Polyp Detection [88.4359020192429]
Existing methods either involve computationally expensive context aggregation or lack prior modeling of polyps, resulting in poor performance in challenging cases.
In this paper, we propose the Enhanced CenterNet with Contrastive Learning (ECC-PolypDet), a two-stage training & end-to-end inference framework.
Box-assisted Contrastive Learning (BCL) during training to minimize the intra-class difference and maximize the inter-class difference between foreground polyps and backgrounds, enabling our model to capture concealed polyps.
In the fine-tuning stage, we introduce the IoU-guided Sample Re-weighting
arXiv Detail & Related papers (2024-01-10T07:03:41Z) - Lesion-aware Dynamic Kernel for Polyp Segmentation [49.63274623103663]
We propose a lesion-aware dynamic network (LDNet) for polyp segmentation.
It is a traditional u-shape encoder-decoder structure incorporated with a dynamic kernel generation and updating scheme.
This simple but effective scheme endows our model with powerful segmentation performance and generalization capability.
arXiv Detail & Related papers (2023-01-12T09:53:57Z) - BoxPolyp:Boost Generalized Polyp Segmentation Using Extra Coarse
Bounding Box Annotations [79.17754846553866]
We propose a boosted BoxPolyp model to make full use of both accurate mask and extra coarse box annotations.
In practice, box annotations are applied to alleviate the over-fitting issue of previous polyp segmentation models.
Our proposed model outperforms previous state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2022-12-07T07:45:50Z) - TGANet: Text-guided attention for improved polyp segmentation [2.3293678240472517]
Colonoscopy is a gold standard procedure but is highly operator-dependent. Automated polyp segmentation, a precancerous precursor, can minimize missed rates and timely treatment of colon cancer at an early stage.
We exploit size-related and polyp number-related features in the form of text attention during training.
Our proposed text-guided attention network (TGANet) can generalize well to variable-sized polyps in different datasets.
arXiv Detail & Related papers (2022-05-09T13:53:26Z) - Stepwise Feature Fusion: Local Guides Global [14.394421688712052]
We propose a new State-Of-The-Art model for medical image segmentation, the SSFormer, which uses a pyramid Transformer encoder to improve the generalization ability of models.
Our proposed Progressive Locality Decoder can be adapted to the pyramid Transformer backbone to emphasize local features and attention dispersion.
arXiv Detail & Related papers (2022-03-07T10:36:38Z) - Automatic Polyp Segmentation via Multi-scale Subtraction Network [100.94922587360871]
In clinical practice, precise polyp segmentation provides important information in the early detection of colorectal cancer.
Most existing methods are based on U-shape structure and use element-wise addition or concatenation to fuse different level features progressively in decoder.
We propose a multi-scale subtraction network (MSNet) to segment polyp from colonoscopy image.
arXiv Detail & Related papers (2021-08-11T07:54:07Z) - DDANet: Dual Decoder Attention Network for Automatic Polyp Segmentation [0.3734402152170273]
We propose a novel architecture called DDANet'' based on a dual decoder attention network.
Experiments demonstrate that the model trained on the Kvasir-SEG dataset and tested on an unseen dataset achieves a dice coefficient of 0.7874, mIoU of 0.7010, recall of 0.7987, and a precision of 0.8577.
arXiv Detail & Related papers (2020-12-30T17:52:35Z) - PraNet: Parallel Reverse Attention Network for Polyp Segmentation [155.93344756264824]
We propose a parallel reverse attention network (PraNet) for accurate polyp segmentation in colonoscopy images.
We first aggregate the features in high-level layers using a parallel partial decoder (PPD)
In addition, we mine the boundary cues using a reverse attention (RA) module, which is able to establish the relationship between areas and boundary cues.
arXiv Detail & Related papers (2020-06-13T08:13:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.