Evaluating Transformer based Semantic Segmentation Networks for
Pathological Image Segmentation
- URL: http://arxiv.org/abs/2108.11993v1
- Date: Thu, 26 Aug 2021 18:46:43 GMT
- Title: Evaluating Transformer based Semantic Segmentation Networks for
Pathological Image Segmentation
- Authors: Cam Nguyen, Zuhayr Asad, Yuankai Huo
- Abstract summary: Histopathology has played an essential role in cancer diagnosis.
Various CNN-based automated pathological image segmentation approaches have been developed in computer-assisted pathological image analysis.
Transformer neural networks (Transformer) have shown the unique merit of capturing the global long distance dependencies across the entire image as a new deep learning paradigm.
- Score: 2.7029872968576947
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Histopathology has played an essential role in cancer diagnosis. With the
rapid advances in convolutional neural networks (CNN). Various CNN-based
automated pathological image segmentation approaches have been developed in
computer-assisted pathological image analysis. In the past few years,
Transformer neural networks (Transformer) have shown the unique merit of
capturing the global long distance dependencies across the entire image as a
new deep learning paradigm. Such merit is appealing for exploring spatially
heterogeneous pathological images. However, there have been very few, if any,
studies that have systematically evaluated the current Transformer based
approaches in pathological image segmentation. To assess the performance of
Transformer segmentation models on whole slide images (WSI), we quantitatively
evaluated six prevalent transformer-based models on tumor segmentation, using
the widely used PAIP liver histopathological dataset. For a more comprehensive
analysis, we also compare the transformer-based models with six major
traditional CNN-based models. The results show that the Transformer-based
models exhibit a general superior performance over the CNN-based models. In
particular, Segmenter, Swin-Transformer and TransUNet, all transformer-based,
came out as the best performers among the twelve evaluated models.
Related papers
- Going Beyond U-Net: Assessing Vision Transformers for Semantic Segmentation in Microscopy Image Analysis [4.151073288078749]
transformer models promise to enhance the segmentation process of microscopy images.
We assess the efficacy of transformers, including UNETR, the Segment Anything Model, and Swin-UPerNet, and compare them with the well-established U-Net model.
The results demonstrate that these modifications improve segmentation performance compared to the classical U-Net model and the unmodified Swin-UPerNet.
arXiv Detail & Related papers (2024-09-25T13:53:48Z) - Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - SeUNet-Trans: A Simple yet Effective UNet-Transformer Model for Medical
Image Segmentation [0.0]
We propose a simple yet effective UNet-Transformer (seUNet-Trans) model for medical image segmentation.
In our approach, the UNet model is designed as a feature extractor to generate multiple feature maps from the input images.
By leveraging the UNet architecture and the self-attention mechanism, our model not only retains the preservation of both local and global context information but also is capable of capturing long-range dependencies between input elements.
arXiv Detail & Related papers (2023-10-16T01:13:38Z) - Breast Ultrasound Tumor Classification Using a Hybrid Multitask
CNN-Transformer Network [63.845552349914186]
Capturing global contextual information plays a critical role in breast ultrasound (BUS) image classification.
Vision Transformers have an improved capability of capturing global contextual information but may distort the local image patterns due to the tokenization operations.
In this study, we proposed a hybrid multitask deep neural network called Hybrid-MT-ESTAN, designed to perform BUS tumor classification and segmentation.
arXiv Detail & Related papers (2023-08-04T01:19:32Z) - Reliable Joint Segmentation of Retinal Edema Lesions in OCT Images [55.83984261827332]
In this paper, we propose a novel reliable multi-scale wavelet-enhanced transformer network.
We develop a novel segmentation backbone that integrates a wavelet-enhanced feature extractor network and a multi-scale transformer module.
Our proposed method achieves better segmentation accuracy with a high degree of reliability as compared to other state-of-the-art segmentation approaches.
arXiv Detail & Related papers (2022-12-01T07:32:56Z) - Transformer based Generative Adversarial Network for Liver Segmentation [4.317557160310758]
We propose a new segmentation approach using a hybrid approach combining the Transformer(s) with the Generative Adversarial Network (GAN) approach.
Our model achieved a high dice coefficient of 0.9433, recall of 0.9515, and precision of 0.9376 and outperformed other Transformer based approaches.
arXiv Detail & Related papers (2022-05-21T19:55:43Z) - From Modern CNNs to Vision Transformers: Assessing the Performance,
Robustness, and Classification Strategies of Deep Learning Models in
Histopathology [1.8947504307591034]
We develop a new methodology to extensively evaluate a wide range of classification models.
We thoroughly tested the models on five widely used histopathology datasets.
We extend existing interpretability methods and systematically reveal insights of the models' classifications strategies.
arXiv Detail & Related papers (2022-04-11T12:26:19Z) - Class-Aware Generative Adversarial Transformers for Medical Image
Segmentation [39.14169989603906]
We present CA-GANformer, a novel type of generative adversarial transformers, for medical image segmentation.
First, we take advantage of the pyramid structure to construct multi-scale representations and handle multi-scale variations.
We then design a novel class-aware transformer module to better learn the discriminative regions of objects with semantic structures.
arXiv Detail & Related papers (2022-01-26T03:50:02Z) - Swin-Unet: Unet-like Pure Transformer for Medical Image Segmentation [63.46694853953092]
Swin-Unet is an Unet-like pure Transformer for medical image segmentation.
tokenized image patches are fed into the Transformer-based U-shaped decoder-Decoder architecture.
arXiv Detail & Related papers (2021-05-12T09:30:26Z) - Medical Transformer: Gated Axial-Attention for Medical Image
Segmentation [73.98974074534497]
We study the feasibility of using Transformer-based network architectures for medical image segmentation tasks.
We propose a Gated Axial-Attention model which extends the existing architectures by introducing an additional control mechanism in the self-attention module.
To train the model effectively on medical images, we propose a Local-Global training strategy (LoGo) which further improves the performance.
arXiv Detail & Related papers (2021-02-21T18:35:14Z) - TransUNet: Transformers Make Strong Encoders for Medical Image
Segmentation [78.01570371790669]
Medical image segmentation is an essential prerequisite for developing healthcare systems.
On various medical image segmentation tasks, the u-shaped architecture, also known as U-Net, has become the de-facto standard.
We propose TransUNet, which merits both Transformers and U-Net, as a strong alternative for medical image segmentation.
arXiv Detail & Related papers (2021-02-08T16:10:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.