Diffusion Probabilistic Models beat GANs on Medical Images
- URL: http://arxiv.org/abs/2212.07501v1
- Date: Wed, 14 Dec 2022 20:46:50 GMT
- Title: Diffusion Probabilistic Models beat GANs on Medical Images
- Authors: Gustav M\"uller-Franzes, Jan Moritz Niehues, Firas Khader, Soroosh
Tayebi Arasteh, Christoph Haarburger, Christiane Kuhl, Tianci Wang, Tianyu
Han, Sven Nebelung, Jakob Nikolas Kather, Daniel Truhn
- Abstract summary: We propose Medfusion, a conditional latent DDPM for medical images.
We compare our DDPM-based model against GAN-based models, which constitute the current state-of-the-art in the medical domain.
Our study shows that DDPM are a superior alternative to GANs for image synthesis in the medical domain.
- Score: 0.13386555802329278
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The success of Deep Learning applications critically depends on the quality
and scale of the underlying training data. Generative adversarial networks
(GANs) can generate arbitrary large datasets, but diversity and fidelity are
limited, which has recently been addressed by denoising diffusion probabilistic
models (DDPMs) whose superiority has been demonstrated on natural images. In
this study, we propose Medfusion, a conditional latent DDPM for medical images.
We compare our DDPM-based model against GAN-based models, which constitute the
current state-of-the-art in the medical domain. Medfusion was trained and
compared with (i) StyleGan-3 on n=101,442 images from the AIROGS challenge
dataset to generate fundoscopies with and without glaucoma, (ii) ProGAN on
n=191,027 from the CheXpert dataset to generate radiographs with and without
cardiomegaly and (iii) wGAN on n=19,557 images from the CRCMS dataset to
generate histopathological images with and without microsatellite stability. In
the AIROGS, CRMCS, and CheXpert datasets, Medfusion achieved lower (=better)
FID than the GANs (11.63 versus 20.43, 30.03 versus 49.26, and 17.28 versus
84.31). Also, fidelity (precision) and diversity (recall) were higher (=better)
for Medfusion in all three datasets. Our study shows that DDPM are a superior
alternative to GANs for image synthesis in the medical domain.
Related papers
- Brain Tumor Classification on MRI in Light of Molecular Markers [61.77272414423481]
Co-deletion of the 1p/19q gene is associated with clinical outcomes in low-grade gliomas.
This study aims to utilize a specially MRI-based convolutional neural network for brain cancer detection.
arXiv Detail & Related papers (2024-09-29T07:04:26Z) - A Domain Translation Framework with an Adversarial Denoising Diffusion
Model to Generate Synthetic Datasets of Echocardiography Images [0.5999777817331317]
We introduce a framework to create echocardiography images suitable to be used for clinical research purposes.
For several domain translation operations, the results verified that such generative model was able to synthesize high quality image samples.
arXiv Detail & Related papers (2024-03-07T15:58:03Z) - GAN-GA: A Generative Model based on Genetic Algorithm for Medical Image
Generation [0.0]
Generative models offer a promising solution for addressing medical image shortage problems.
This paper proposes the GAN-GA, a generative model optimized by embedding a genetic algorithm.
The proposed model enhances image fidelity and diversity while preserving distinctive features.
arXiv Detail & Related papers (2023-12-30T20:16:45Z) - EMIT-Diff: Enhancing Medical Image Segmentation via Text-Guided
Diffusion Model [4.057796755073023]
We develop controllable diffusion models for medical image synthesis, called EMIT-Diff.
We leverage recent diffusion probabilistic models to generate realistic and diverse synthetic medical image data.
In our approach, we ensure that the synthesized samples adhere to medically relevant constraints.
arXiv Detail & Related papers (2023-10-19T16:18:02Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - Conditional Diffusion Models for Semantic 3D Brain MRI Synthesis [0.0]
We introduce Med-DDPM, a diffusion model designed for 3D semantic brain MRI synthesis.
It effectively tackles data scarcity and privacy issues by integrating semantic conditioning.
It generates diverse, coherent images with high visual fidelity.
arXiv Detail & Related papers (2023-05-29T04:14:38Z) - Generative models improve fairness of medical classifiers under
distribution shifts [49.10233060774818]
We show that learning realistic augmentations automatically from data is possible in a label-efficient manner using generative models.
We demonstrate that these learned augmentations can surpass ones by making models more robust and statistically fair in- and out-of-distribution.
arXiv Detail & Related papers (2023-04-18T18:15:38Z) - SAG-GAN: Semi-Supervised Attention-Guided GANs for Data Augmentation on
Medical Images [47.35184075381965]
We present a data augmentation method for generating synthetic medical images using cycle-consistency Generative Adversarial Networks (GANs)
The proposed GANs-based model can generate a tumor image from a normal image, and in turn, it can also generate a normal image from a tumor image.
We train the classification model using real images with classic data augmentation methods and classification models using synthetic images.
arXiv Detail & Related papers (2020-11-15T14:01:24Z) - Fader Networks for domain adaptation on fMRI: ABIDE-II study [68.5481471934606]
We use 3D convolutional autoencoders to build the domain irrelevant latent space image representation and demonstrate this method to outperform existing approaches on ABIDE data.
arXiv Detail & Related papers (2020-10-14T16:50:50Z) - Classification of COVID-19 in CT Scans using Multi-Source Transfer
Learning [91.3755431537592]
We propose the use of Multi-Source Transfer Learning to improve upon traditional Transfer Learning for the classification of COVID-19 from CT scans.
With our multi-source fine-tuning approach, our models outperformed baseline models fine-tuned with ImageNet.
Our best performing model was able to achieve an accuracy of 0.893 and a Recall score of 0.897, outperforming its baseline Recall score by 9.3%.
arXiv Detail & Related papers (2020-09-22T11:53:06Z) - Automated Chest CT Image Segmentation of COVID-19 Lung Infection based
on 3D U-Net [0.0]
The coronavirus disease 2019 (COVID-19) affects billions of lives around the world and has a significant impact on public healthcare.
We propose an innovative automated segmentation pipeline for COVID-19 infected regions.
Our method focuses on on-the-fly generation of unique and random image patches for training by performing several preprocessing methods.
arXiv Detail & Related papers (2020-06-24T17:29:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.