TFCNs: A CNN-Transformer Hybrid Network for Medical Image Segmentation
- URL: http://arxiv.org/abs/2207.03450v1
- Date: Thu, 7 Jul 2022 17:28:18 GMT
- Title: TFCNs: A CNN-Transformer Hybrid Network for Medical Image Segmentation
- Authors: Zihan Li, Dihan Li, Cangbai Xu, Weice Wang, Qingqi Hong, Qingde Li,
Jie Tian
- Abstract summary: We propose TFCNs (Transformers for Fully Convolutional denseNets) to tackle the problem of medical image segmentation.
We show that TFCNs can achieve state-of-the-art performance with dice scores of 83.72% on the Synapse dataset.
- Score: 4.768284586442054
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical image segmentation is one of the most fundamental tasks concerning
medical information analysis. Various solutions have been proposed so far,
including many deep learning-based techniques, such as U-Net, FC-DenseNet, etc.
However, high-precision medical image segmentation remains a highly challenging
task due to the existence of inherent magnification and distortion in medical
images as well as the presence of lesions with similar density to normal
tissues. In this paper, we propose TFCNs (Transformers for Fully Convolutional
denseNets) to tackle the problem by introducing ResLinear-Transformer
(RL-Transformer) and Convolutional Linear Attention Block (CLAB) to
FC-DenseNet. TFCNs is not only able to utilize more latent information from the
CT images for feature extraction, but also can capture and disseminate semantic
features and filter non-semantic features more effectively through the CLAB
module. Our experimental results show that TFCNs can achieve state-of-the-art
performance with dice scores of 83.72\% on the Synapse dataset. In addition, we
evaluate the robustness of TFCNs for lesion area effects on the COVID-19 public
datasets. The Python code will be made publicly available on
https://github.com/HUANGLIZI/TFCNs.
Related papers
- TransResNet: Integrating the Strengths of ViTs and CNNs for High Resolution Medical Image Segmentation via Feature Grafting [6.987177704136503]
High-resolution images are preferable in medical imaging domain as they significantly improve the diagnostic capability of the underlying method.
Most of the existing deep learning-based techniques for medical image segmentation are optimized for input images having small spatial dimensions and perform poorly on high-resolution images.
We propose a parallel-in-branch architecture called TransResNet, which incorporates Transformer and CNN in a parallel manner to extract features from multi-resolution images independently.
arXiv Detail & Related papers (2024-10-01T18:22:34Z) - BEFUnet: A Hybrid CNN-Transformer Architecture for Precise Medical Image
Segmentation [0.0]
This paper proposes an innovative U-shaped network called BEFUnet, which enhances the fusion of body and edge information for precise medical image segmentation.
The BEFUnet comprises three main modules, including a novel Local Cross-Attention Feature (LCAF) fusion module, a novel Double-Level Fusion (DLF) module, and dual-branch encoder.
The LCAF module efficiently fuses edge and body features by selectively performing local cross-attention on features that are spatially close between the two modalities.
arXiv Detail & Related papers (2024-02-13T21:03:36Z) - Reliable Joint Segmentation of Retinal Edema Lesions in OCT Images [55.83984261827332]
In this paper, we propose a novel reliable multi-scale wavelet-enhanced transformer network.
We develop a novel segmentation backbone that integrates a wavelet-enhanced feature extractor network and a multi-scale transformer module.
Our proposed method achieves better segmentation accuracy with a high degree of reliability as compared to other state-of-the-art segmentation approaches.
arXiv Detail & Related papers (2022-12-01T07:32:56Z) - HistoSeg : Quick attention with multi-loss function for multi-structure
segmentation in digital histology images [0.696194614504832]
Medical image segmentation assists in computer-aided diagnosis, surgeries, and treatment.
We proposed an generalization-Decoder Network, Quick Attention Module and a Multi Loss Function.
We evaluate the capability of our proposed network on two publicly available datasets for medical image segmentation MoNuSeg and GlaS.
arXiv Detail & Related papers (2022-09-01T21:10:00Z) - Data-Efficient Vision Transformers for Multi-Label Disease
Classification on Chest Radiographs [55.78588835407174]
Vision Transformers (ViTs) have not been applied to this task despite their high classification performance on generic images.
ViTs do not rely on convolutions but on patch-based self-attention and in contrast to CNNs, no prior knowledge of local connectivity is present.
Our results show that while the performance between ViTs and CNNs is on par with a small benefit for ViTs, DeiTs outperform the former if a reasonably large data set is available for training.
arXiv Detail & Related papers (2022-08-17T09:07:45Z) - Preservation of High Frequency Content for Deep Learning-Based Medical
Image Classification [74.84221280249876]
An efficient analysis of large amounts of chest radiographs can aid physicians and radiologists.
We propose a novel Discrete Wavelet Transform (DWT)-based method for the efficient identification and encoding of visual information.
arXiv Detail & Related papers (2022-05-08T15:29:54Z) - LEFM-Nets: Learnable Explicit Feature Map Deep Networks for Segmentation
of Histopathological Images of Frozen Sections [0.10152838128195464]
We propose a framework that embeds existing DNs into a low-dimensional subspace induced by the learnable explicit feature map layer.
New LEFM-Nets are applied to the segmentation of adenocarcinoma of a colon in a liver from images of hematoxylin and eosin stained frozen sections.
arXiv Detail & Related papers (2022-04-14T13:27:45Z) - InDuDoNet+: A Model-Driven Interpretable Dual Domain Network for Metal
Artifact Reduction in CT Images [53.4351366246531]
We construct a novel interpretable dual domain network, termed InDuDoNet+, into which CT imaging process is finely embedded.
We analyze the CT values among different tissues, and merge the prior observations into a prior network for our InDuDoNet+, which significantly improve its generalization performance.
arXiv Detail & Related papers (2021-12-23T15:52:37Z) - Transformer-Unet: Raw Image Processing with Unet [4.7944896477309555]
We propose Transformer-Unet by adding transformer modules in raw images instead of feature maps in Unet.
We form an end-to-end network and gain segmentation results better than many previous Unet based algorithms in our experiment.
arXiv Detail & Related papers (2021-09-17T09:03:10Z) - CoTr: Efficiently Bridging CNN and Transformer for 3D Medical Image
Segmentation [95.51455777713092]
Convolutional neural networks (CNNs) have been the de facto standard for nowadays 3D medical image segmentation.
We propose a novel framework that efficiently bridges a bf Convolutional neural network and a bf Transformer bf (CoTr) for accurate 3D medical image segmentation.
arXiv Detail & Related papers (2021-03-04T13:34:22Z) - TransUNet: Transformers Make Strong Encoders for Medical Image
Segmentation [78.01570371790669]
Medical image segmentation is an essential prerequisite for developing healthcare systems.
On various medical image segmentation tasks, the u-shaped architecture, also known as U-Net, has become the de-facto standard.
We propose TransUNet, which merits both Transformers and U-Net, as a strong alternative for medical image segmentation.
arXiv Detail & Related papers (2021-02-08T16:10:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.