Expert Validation of Synthetic Cervical Spine Radiographs Generated with a Denoising Diffusion Probabilistic Model
- URL: http://arxiv.org/abs/2510.22166v1
- Date: Sat, 25 Oct 2025 05:25:37 GMT
- Title: Expert Validation of Synthetic Cervical Spine Radiographs Generated with a Denoising Diffusion Probabilistic Model
- Authors: Austin A. Barr, Brij S. Karmur, Anthony J. Winder, Eddie Guo, John T. Lysack, James N. Scott, William F. Morrish, Muneer Eesa, Morgan Willson, David W. Cadotte, Michael M. H. Yang, Ian Y. M. Chan, Sanju Lama, Garnette R. Sutherland,
- Abstract summary: We generate realistic cervical spine radiographs using a denoising diffusion probabilistic model (DDPM) trained on 4,963 images from the Cervical Spine X-ray Atlas.<n>Experts reviewed 50 quartets containing one real and three synthetic images, identifying the real image rating realism on a 4-pointrt scale.<n>These results demonstrate that DDPM-generated cervical spine X-rays are statistically indistinguishable in realism and quality from real clinical images.
- Score: 0.6873565341012776
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Machine learning in neurosurgery is limited by challenges in assembling large, high-quality imaging datasets. Synthetic data offers a scalable, privacy-preserving solution. We evaluated the feasibility of generating realistic lateral cervical spine radiographs using a denoising diffusion probabilistic model (DDPM) trained on 4,963 images from the Cervical Spine X-ray Atlas. Model performance was monitored via training/validation loss and Frechet inception distance, and synthetic image quality was assessed in a blinded "clinical Turing test" with six neuroradiologists and two spine-fellowship trained neurosurgeons. Experts reviewed 50 quartets containing one real and three synthetic images, identifying the real image and rating realism on a 4-point Likert scale. Experts correctly identified the real image in 29% of trials (Fleiss' kappa=0.061). Mean realism scores were comparable between real (3.323) and synthetic images (3.228, 3.258, and 3.320; p=0.383, 0.471, 1.000). Nearest-neighbor analysis found no evidence of memorization. We also provide a dataset of 20,063 synthetic radiographs. These results demonstrate that DDPM-generated cervical spine X-rays are statistically indistinguishable in realism and quality from real clinical images, offering a novel approach to creating large-scale neuroimaging datasets for ML applications in landmarking, segmentation, and classification.
Related papers
- Bridging the Gap Between Ideal and Real-world Evaluation: Benchmarking AI-Generated Image Detection in Challenging Scenarios [54.07895223545793]
This paper introduces the Real-World Robustness dataset (RRDataset) for comprehensive evaluation of detection models across three dimensions.<n>RRDataset includes high-quality images from seven major scenarios.<n>We benchmarked 17 detectors and 10 vision-language models (VLMs) on RRDataset and conducted a large-scale human study.
arXiv Detail & Related papers (2025-09-11T06:15:52Z) - PanoDiff-SR: Synthesizing Dental Panoramic Radiographs using Diffusion and Super-resolution [60.970656010712275]
We propose a combination of diffusion-based generation (PanoDiff) and Super-Resolution (SR) for generating synthetic dental panoramic radiographs (PRs)<n>The former generates a low-resolution (LR) seed of a PR which is then processed by the SR model to yield a high-resolution (HR) PR of size 1024 X 512.<n>For SR, we propose a state-of-the-art transformer that learns local-global relationships, resulting in sharper edges and textures.
arXiv Detail & Related papers (2025-07-12T09:52:10Z) - Automated MRI Quality Assessment of Brain T1-weighted MRI in Clinical Data Warehouses: A Transfer Learning Approach Relying on Artefact Simulation [3.115212915804253]
This study presents an innovative transfer learning method for automated quality control of 3D gradient echo T1-weighted brain MRIs within a clinical data warehouse.
We first intentionally corrupt images from research datasets by inducing poorer contrast, adding noise and introducing motion artefacts.
Three artefact-specific models are pre-trained using these corrupted images to detect distinct types of artefacts.
arXiv Detail & Related papers (2024-06-18T09:53:07Z) - Robust and Interpretable COVID-19 Diagnosis on Chest X-ray Images using
Adversarial Training [0.8287206589886881]
We train 21 convolutional neural network (CNN) models on a diverse set of 33,000+ chest X-ray (CXR) images to classify between healthy, COVID-19, and non-COVID-19 pneumonia CXRs.
Our resulting models achieved a 3-way classification accuracy, recall, and precision of up to 97.03%, 97.97%, and 99.95%, respectively.
arXiv Detail & Related papers (2023-11-23T23:40:01Z) - The Impact of Loss Functions and Scene Representations for 3D/2D
Registration on Single-view Fluoroscopic X-ray Pose Estimation [1.758213853394712]
We first develop a differentiable projection rendering framework for the efficient computation of Digitally Reconstructed Radiographs (DRRs)
We then perform pose estimation by iterative descent using various candidate loss functions, that quantify the image discrepancy of the synthesized DRR with respect to the ground-truth fluoroscopic X-ray image.
Using the Mutual Information loss, a comprehensive evaluation of pose estimation performed on a tomographic X-ray dataset of 50 patients$'$ skulls shows that utilizing either discretized (CBCT) or neural (NeTT/mNeRF) scene representations in DiffProj leads to
arXiv Detail & Related papers (2023-08-01T01:12:29Z) - Bi-parametric prostate MR image synthesis using pathology and
sequence-conditioned stable diffusion [3.290987481767681]
We propose an image synthesis mechanism for multi-sequence prostate MR images conditioned on text.
We generate paired bi-parametric images conditioned on images conditioned on paired data.
We validate our method using 2D image slices from real suspected prostate cancer patients.
arXiv Detail & Related papers (2023-03-03T17:24:39Z) - Significantly improving zero-shot X-ray pathology classification via fine-tuning pre-trained image-text encoders [50.689585476660554]
We propose a new fine-tuning strategy that includes positive-pair loss relaxation and random sentence sampling.
Our approach consistently improves overall zero-shot pathology classification across four chest X-ray datasets and three pre-trained models.
arXiv Detail & Related papers (2022-12-14T06:04:18Z) - Medical Diffusion -- Denoising Diffusion Probabilistic Models for 3D
Medical Image Generation [0.6486409713123691]
We show that diffusion probabilistic models can synthesize high quality medical imaging data.
We provide quantitative measurements of their performance through a reader study with two medical experts.
We demonstrate that synthetic images can be used in a self-supervised pre-training and improve the performance of breast segmentation models when data is scarce.
arXiv Detail & Related papers (2022-11-07T08:37:48Z) - Automated SSIM Regression for Detection and Quantification of Motion
Artefacts in Brain MR Images [54.739076152240024]
Motion artefacts in magnetic resonance brain images are a crucial issue.
The assessment of MR image quality is fundamental before proceeding with the clinical diagnosis.
An automated image quality assessment based on the structural similarity index (SSIM) regression has been proposed here.
arXiv Detail & Related papers (2022-06-14T10:16:54Z) - Towards Ultrafast MRI via Extreme k-Space Undersampling and
Superresolution [65.25508348574974]
We go below the MRI acceleration factors reported by all published papers that reference the original fastMRI challenge.
We consider powerful deep learning based image enhancement methods to compensate for the underresolved images.
The quality of the reconstructed images surpasses that of the other methods, yielding an MSE of 0.00114, a PSNR of 29.6 dB, and an SSIM of 0.956 at x16 acceleration factor.
arXiv Detail & Related papers (2021-03-04T10:45:01Z) - XraySyn: Realistic View Synthesis From a Single Radiograph Through CT
Priors [118.27130593216096]
A radiograph visualizes the internal anatomy of a patient through the use of X-ray, which projects 3D information onto a 2D plane.
To the best of our knowledge, this is the first work on radiograph view synthesis.
We show that by gaining an understanding of radiography in 3D space, our method can be applied to radiograph bone extraction and suppression without groundtruth bone labels.
arXiv Detail & Related papers (2020-12-04T05:08:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.