CTformer: Convolution-free Token2Token Dilated Vision Transformer for
Low-dose CT Denoising
- URL: http://arxiv.org/abs/2202.13517v1
- Date: Mon, 28 Feb 2022 02:58:16 GMT
- Title: CTformer: Convolution-free Token2Token Dilated Vision Transformer for
Low-dose CT Denoising
- Authors: Dayang Wang, Fenglei Fan, Zhan Wu, Rui Liu, Fei Wang, Hengyong Yu
- Abstract summary: Low-dose computed tomography (LDCT) denoising is an important problem in CT research.
vision transformers have shown superior feature representation ability over convolutional neural networks (CNNs)
We propose a Convolution-free Token2Token Dilated Vision Transformer for low-dose CT denoising.
- Score: 11.67382017798666
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Low-dose computed tomography (LDCT) denoising is an important problem in CT
research. Compared to the normal dose CT (NDCT), LDCT images are subjected to
severe noise and artifacts. Recently in many studies, vision transformers have
shown superior feature representation ability over convolutional neural
networks (CNNs). However, unlike CNNs, the potential of vision transformers in
LDCT denoising was little explored so far. To fill this gap, we propose a
Convolution-free Token2Token Dilated Vision Transformer for low-dose CT
denoising. The CTformer uses a more powerful token rearrangement to encompass
local contextual information and thus avoids convolution. It also dilates and
shifts feature maps to capture longer-range interaction. We interpret the
CTformer by statically inspecting patterns of its internal attention maps and
dynamically tracing the hierarchical attention flow with an explanatory graph.
Furthermore, an overlapped inference mechanism is introduced to effectively
eliminate the boundary artifacts that are common for encoder-decoder-based
denoising models. Experimental results on Mayo LDCT dataset suggest that the
CTformer outperforms the state-of-the-art denoising methods with a low
computation overhead.
Related papers
- WIA-LD2ND: Wavelet-based Image Alignment for Self-supervised Low-Dose CT Denoising [74.14134385961775]
We introduce a novel self-supervised CT image denoising method called WIA-LD2ND, only using NDCT data.
WIA-LD2ND comprises two modules: Wavelet-based Image Alignment (WIA) and Frequency-Aware Multi-scale Loss (FAM)
arXiv Detail & Related papers (2024-03-18T11:20:11Z) - Rotational Augmented Noise2Inverse for Low-dose Computed Tomography
Reconstruction [83.73429628413773]
Supervised deep learning methods have shown the ability to remove noise in images but require accurate ground truth.
We propose a novel self-supervised framework for LDCT, in which ground truth is not required for training the convolutional neural network (CNN)
Numerical and experimental results show that the reconstruction accuracy of N2I with sparse views is degrading while the proposed rotational augmented Noise2Inverse (RAN2I) method keeps better image quality over a different range of sampling angles.
arXiv Detail & Related papers (2023-12-19T22:40:51Z) - Deep learning network to correct axial and coronal eye motion in 3D OCT
retinal imaging [65.47834983591957]
We propose deep learning based neural networks to correct axial and coronal motion artifacts in OCT based on a single scan.
The experimental result shows that the proposed method can effectively correct motion artifacts and achieve smaller error than other methods.
arXiv Detail & Related papers (2023-05-27T03:55:19Z) - The role of noise in denoising models for anomaly detection in medical
images [62.0532151156057]
Pathological brain lesions exhibit diverse appearance in brain images.
Unsupervised anomaly detection approaches have been proposed using only normal data for training.
We show that optimization of the spatial resolution and magnitude of the noise improves the performance of different model training regimes.
arXiv Detail & Related papers (2023-01-19T21:39:38Z) - Self Supervised Low Dose Computed Tomography Image Denoising Using
Invertible Network Exploiting Inter Slice Congruence [20.965610734723636]
This study proposes a novel method for self-supervised low-dose CT denoising to alleviate the requirement of paired LDCT and NDCT images.
We have trained an invertible neural network to minimize the pixel-based mean square distance between a noisy slice and the average of its two immediate adjacent noisy slices.
arXiv Detail & Related papers (2022-11-03T07:16:18Z) - Self-supervised Physics-based Denoising for Computed Tomography [2.2758845733923687]
Computed Tomography (CT) imposes risk on the patients due to its inherent X-ray radiation.
Lowering the radiation dose reduces the health risks but leads to noisier measurements, which decreases the tissue contrast and causes artifacts in CT images.
Modern deep learning noise suppression methods alleviate the challenge but require low-noise-high-noise CT image pairs for training.
We introduce a new self-supervised approach for CT denoising Noise2NoiseTD-ANM that can be trained without the high-dose CT projection ground truth images.
arXiv Detail & Related papers (2022-11-01T20:58:50Z) - Masked Autoencoders for Low dose CT denoising [9.575051352192697]
Masked autoencoders (MAE) have been proposed as an effective label-free self-pretraining method for transformers.
We redesign the classical encoder-decoder learning model to match the denoising task and apply it to LDCT denoising problem.
arXiv Detail & Related papers (2022-10-10T18:27:58Z) - Multi-stage image denoising with the wavelet transform [125.2251438120701]
Deep convolutional neural networks (CNNs) are used for image denoising via automatically mining accurate structure information.
We propose a multi-stage image denoising CNN with the wavelet transform (MWDCNN) via three stages, i.e., a dynamic convolutional block (DCB), two cascaded wavelet transform and enhancement blocks (WEBs) and residual block (RB)
arXiv Detail & Related papers (2022-09-26T03:28:23Z) - Patch-wise Deep Metric Learning for Unsupervised Low-Dose CT Denoising [33.706959549595496]
We propose a novel unsupervised learning approach for lowdose CT reconstruction using patch-wise deep metric learning.
The key idea is to learn embedding space by pulling the positive pairs of image patches which shares the same anatomical structure, and pushing the negative pairs which have same noise level each other.
Experimental results confirm that our deep metric learning plays a critical role in producing high quality denoised images without CT number shift.
arXiv Detail & Related papers (2022-07-06T00:58:11Z) - CNN Filter Learning from Drawn Markers for the Detection of Suggestive
Signs of COVID-19 in CT Images [58.720142291102135]
We propose a method that does not require either large annotated datasets or backpropagation to estimate the filters of a convolutional neural network (CNN)
For a few CT images, the user draws markers at representative normal and abnormal regions.
The method generates a feature extractor composed of a sequence of convolutional layers, whose kernels are specialized in enhancing regions similar to the marked ones.
arXiv Detail & Related papers (2021-11-16T15:03:42Z) - TED-net: Convolution-free T2T Vision Transformer-based Encoder-decoder
Dilation network for Low-dose CT Denoising [5.2227817530931535]
We propose a convolution-free vision transformer-based-decoder Dilation net-work (TED-net) to enrich the family of LDCT denoising algorithms.
Our model is evaluated on the AAPM-Mayo clinic LDCT Grand Challenge dataset, and results show outperformance over the state-of-the-art denoising methods.
arXiv Detail & Related papers (2021-06-08T19:26:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.