LoMAE: Low-level Vision Masked Autoencoders for Low-dose CT Denoising
- URL: http://arxiv.org/abs/2310.12405v1
- Date: Thu, 19 Oct 2023 01:34:30 GMT
- Title: LoMAE: Low-level Vision Masked Autoencoders for Low-dose CT Denoising
- Authors: Dayang Wang, Yongshun Xu, Shuo Han, Zhan Wu, Li Zhou, Bahareh
Morovati, Hengyong Yu
- Abstract summary: Masked autoencoders (MAE) have been recognized as an effective label-free self-pretraining method for transformers.
We introduce an MAE-GradCAM method to shed light on the latent learning mechanisms of the MAE/LoMAE.
Experiments show that the proposed LoMAE can enhance the transformer's denoising performance and greatly relieve the dependence on the ground truth clean data.
- Score: 5.251624007533231
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Low-dose computed tomography (LDCT) offers reduced X-ray radiation exposure
but at the cost of compromised image quality, characterized by increased noise
and artifacts. Recently, transformer models emerged as a promising avenue to
enhance LDCT image quality. However, the success of such models relies on a
large amount of paired noisy and clean images, which are often scarce in
clinical settings. In the fields of computer vision and natural language
processing, masked autoencoders (MAE) have been recognized as an effective
label-free self-pretraining method for transformers, due to their exceptional
feature representation ability. However, the original pretraining and
fine-tuning design fails to work in low-level vision tasks like denoising. In
response to this challenge, we redesign the classical encoder-decoder learning
model and facilitate a simple yet effective low-level vision MAE, referred to
as LoMAE, tailored to address the LDCT denoising problem. Moreover, we
introduce an MAE-GradCAM method to shed light on the latent learning mechanisms
of the MAE/LoMAE. Additionally, we explore the LoMAE's robustness and
generability across a variety of noise levels. Experiments results show that
the proposed LoMAE can enhance the transformer's denoising performance and
greatly relieve the dependence on the ground truth clean data. It also
demonstrates remarkable robustness and generalizability over a spectrum of
noise levels.
Related papers
- DenoMamba: A fused state-space model for low-dose CT denoising [6.468495781611433]
Low-dose computed tomography (LDCT) lower potential risks linked to radiation exposure.
LDCT denoising is based on neural network models that learn data-driven image priors to separate noise evoked by dose reduction from underlying tissue signals.
DenoMamba is a novel denoising method based on state-space modeling (SSM) that efficiently captures short- and long-range context in medical images.
arXiv Detail & Related papers (2024-09-19T21:32:07Z) - Disruptive Autoencoders: Leveraging Low-level features for 3D Medical
Image Pre-training [51.16994853817024]
This work focuses on designing an effective pre-training framework for 3D radiology images.
We introduce Disruptive Autoencoders, a pre-training framework that attempts to reconstruct the original image from disruptions created by a combination of local masking and low-level perturbations.
The proposed pre-training framework is tested across multiple downstream tasks and achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-07-31T17:59:42Z) - Denoising Simulated Low-Field MRI (70mT) using Denoising Autoencoders
(DAE) and Cycle-Consistent Generative Adversarial Networks (Cycle-GAN) [68.8204255655161]
Cycle Consistent Generative Adversarial Network (GAN) is implemented to yield high-field, high resolution, high signal-to-noise ratio (SNR) Magnetic Resonance Imaging (MRI) images.
Images were utilized to train a Denoising Autoencoder (DAE) and a Cycle-GAN, with paired and unpaired cases.
This work demonstrates the use of a generative deep learning model that can outperform classical DAEs to improve low-field MRI images and does not require image pairs.
arXiv Detail & Related papers (2023-07-12T00:01:00Z) - Stimulating the Diffusion Model for Image Denoising via Adaptive Embedding and Ensembling [56.506240377714754]
We present a novel strategy called the Diffusion Model for Image Denoising (DMID)
Our strategy includes an adaptive embedding method that embeds the noisy image into a pre-trained unconditional diffusion model.
Our DMID strategy achieves state-of-the-art performance on both distortion-based and perception-based metrics.
arXiv Detail & Related papers (2023-07-08T14:59:41Z) - On Sensitivity and Robustness of Normalization Schemes to Input
Distribution Shifts in Automatic MR Image Diagnosis [58.634791552376235]
Deep Learning (DL) models have achieved state-of-the-art performance in diagnosing multiple diseases using reconstructed images as input.
DL models are sensitive to varying artifacts as it leads to changes in the input data distribution between the training and testing phases.
We propose to use other normalization techniques, such as Group Normalization and Layer Normalization, to inject robustness into model performance against varying image artifacts.
arXiv Detail & Related papers (2023-06-23T03:09:03Z) - Masked Autoencoders for Low dose CT denoising [9.575051352192697]
Masked autoencoders (MAE) have been proposed as an effective label-free self-pretraining method for transformers.
We redesign the classical encoder-decoder learning model to match the denoising task and apply it to LDCT denoising problem.
arXiv Detail & Related papers (2022-10-10T18:27:58Z) - TED-net: Convolution-free T2T Vision Transformer-based Encoder-decoder
Dilation network for Low-dose CT Denoising [5.2227817530931535]
We propose a convolution-free vision transformer-based-decoder Dilation net-work (TED-net) to enrich the family of LDCT denoising algorithms.
Our model is evaluated on the AAPM-Mayo clinic LDCT Grand Challenge dataset, and results show outperformance over the state-of-the-art denoising methods.
arXiv Detail & Related papers (2021-06-08T19:26:55Z) - Noise Conscious Training of Non Local Neural Network powered by Self
Attentive Spectral Normalized Markovian Patch GAN for Low Dose CT Denoising [20.965610734723636]
Deep learning-based technique has emerged as a dominant method for low dose CT(LDCT) denoising.
In this study, we propose a novel convolutional module as the first attempt to utilize neighborhood similarity of CT images for denoising tasks.
Next, we moved towards the problem of non-stationarity of CT noise and introduced a new noise aware mean square error loss for LDCT denoising.
arXiv Detail & Related papers (2020-11-11T10:44:52Z) - Progressive Training of Multi-level Wavelet Residual Networks for Image
Denoising [80.10533234415237]
This paper presents a multi-level wavelet residual network (MWRN) architecture as well as a progressive training scheme to improve image denoising performance.
Experiments on both synthetic and real-world noisy images show that our PT-MWRN performs favorably against the state-of-the-art denoising methods.
arXiv Detail & Related papers (2020-10-23T14:14:00Z) - Improving Blind Spot Denoising for Microscopy [73.94017852757413]
We present a novel way to improve the quality of self-supervised denoising.
We assume the clean image to be the result of a convolution with a point spread function (PSF) and explicitly include this operation at the end of our neural network.
arXiv Detail & Related papers (2020-08-19T13:06:24Z) - Self-supervised Denoising via Diffeomorphic Template Estimation:
Application to Optical Coherence Tomography [6.197149831796131]
We propose a joint diffeomorphic template estimation and denoising framework which enables the use of self-supervised denoising for motion deformed repeat acquisitions.
Strong qualitative and quantitative improvements are achieved in denoising OCT images, with generic utility in any imaging modality amenable to multiple exposures.
arXiv Detail & Related papers (2020-08-18T16:52:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.