On the notion of Hallucinations from the lens of Bias and Validity in
Synthetic CXR Images
- URL: http://arxiv.org/abs/2312.06979v1
- Date: Tue, 12 Dec 2023 04:41:20 GMT
- Title: On the notion of Hallucinations from the lens of Bias and Validity in
Synthetic CXR Images
- Authors: Gauri Bhardwaj, Yuvaraj Govindarajulu, Sundaraparipurnan Narayanan,
Pavan Kulkarni, Manojkumar Parmar
- Abstract summary: Generative models, such as diffusion models, aim to mitigate data quality and clinical information disparities.
At Stanford, researchers explored the utility of a fine-tuned Stable Diffusion model (RoentGen) for medical imaging data augmentation.
We leveraged RoentGen to produce synthetic Chest-XRay (CXR) images and conducted assessments on bias, validity, and hallucinations.
- Score: 0.35998666903987897
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Medical imaging has revolutionized disease diagnosis, yet the potential is
hampered by limited access to diverse and privacy-conscious datasets.
Open-source medical datasets, while valuable, suffer from data quality and
clinical information disparities. Generative models, such as diffusion models,
aim to mitigate these challenges. At Stanford, researchers explored the utility
of a fine-tuned Stable Diffusion model (RoentGen) for medical imaging data
augmentation. Our work examines specific considerations to expand the Stanford
research question, Could Stable Diffusion Solve a Gap in Medical Imaging Data?
from the lens of bias and validity of the generated outcomes. We leveraged
RoentGen to produce synthetic Chest-XRay (CXR) images and conducted assessments
on bias, validity, and hallucinations. Diagnostic accuracy was evaluated by a
disease classifier, while a COVID classifier uncovered latent hallucinations.
The bias analysis unveiled disparities in classification performance among
various subgroups, with a pronounced impact on the Female Hispanic subgroup.
Furthermore, incorporating race and gender into input prompts exacerbated
fairness issues in the generated images. The quality of synthetic images
exhibited variability, particularly in certain disease classes, where there was
more significant uncertainty compared to the original images. Additionally, we
observed latent hallucinations, with approximately 42% of the images
incorrectly indicating COVID, hinting at the presence of hallucinatory
elements. These identifications provide new research directions towards
interpretability of synthetic CXR images, for further understanding of
associated risks and patient safety in medical applications.
Related papers
- FairSkin: Fair Diffusion for Skin Disease Image Generation [54.29840149709033]
Diffusion Model (DM) has become a leading method in generating synthetic medical images, but it suffers from a critical twofold bias.
We propose FairSkin, a novel DM framework that mitigates these biases through a three-level resampling mechanism.
Our approach significantly improves the diversity and quality of generated images, contributing to more equitable skin disease detection in clinical settings.
arXiv Detail & Related papers (2024-10-29T21:37:03Z) - Diffusion Facial Forgery Detection [56.69763252655695]
This paper introduces DiFF, a comprehensive dataset dedicated to face-focused diffusion-generated images.
We conduct extensive experiments on the DiFF dataset via a human test and several representative forgery detection methods.
The results demonstrate that the binary detection accuracy of both human observers and automated detectors often falls below 30%.
arXiv Detail & Related papers (2024-01-29T03:20:19Z) - VALD-MD: Visual Attribution via Latent Diffusion for Medical Diagnostics [0.0]
Visual attribution in medical imaging seeks to make evident the diagnostically-relevant components of a medical image.
We here present a novel generative visual attribution technique, one that leverages latent diffusion models in combination with domain-specific large language models.
The resulting system also exhibits a range of latent capabilities including zero-shot localized disease induction.
arXiv Detail & Related papers (2024-01-02T19:51:49Z) - Counterfactual Image Synthesis for Discovery of Personalized Predictive
Image Markers [0.293168019422713]
We show how a deep conditional generative model can be used to perturb local imaging features in baseline images that are pertinent to subject-specific future disease evolution.
Our model produces counterfactuals with changes in imaging features that reflect established clinical markers predictive of future MRI lesional activity at the population level.
arXiv Detail & Related papers (2022-08-03T18:58:45Z) - Texture Characterization of Histopathologic Images Using Ecological
Diversity Measures and Discrete Wavelet Transform [82.53597363161228]
This paper proposes a method for characterizing texture across histopathologic images with a considerable success rate.
It is possible to quantify the intrinsic properties of such images with promising accuracy on two HI datasets.
arXiv Detail & Related papers (2022-02-27T02:19:09Z) - Malignancy Prediction and Lesion Identification from Clinical
Dermatological Images [65.1629311281062]
We consider machine-learning-based malignancy prediction and lesion identification from clinical dermatological images.
We first identify all lesions present in the image regardless of sub-type or likelihood of malignancy, then it estimates their likelihood of malignancy, and through aggregation, it also generates an image-level likelihood of malignancy.
arXiv Detail & Related papers (2021-04-02T20:52:05Z) - Variational Knowledge Distillation for Disease Classification in Chest
X-Rays [102.04931207504173]
We propose itvariational knowledge distillation (VKD), which is a new probabilistic inference framework for disease classification based on X-rays.
We demonstrate the effectiveness of our method on three public benchmark datasets with paired X-ray images and EHRs.
arXiv Detail & Related papers (2021-03-19T14:13:56Z) - Overcoming Barriers to Data Sharing with Medical Image Generation: A
Comprehensive Evaluation [17.983449515155414]
We utilize Generative Adversarial Networks (GANs) to create derived medical imaging datasets consisting entirely of synthetic patient data.
The synthetic images ideally have, in aggregate, similar statistical properties to those of a source dataset but do not contain sensitive personal information.
We measure the synthetic image quality by the performance difference of predictive models trained on either the synthetic or the real dataset.
arXiv Detail & Related papers (2020-11-29T15:41:46Z) - Semi-supervised Medical Image Classification with Relation-driven
Self-ensembling Model [71.80319052891817]
We present a relation-driven semi-supervised framework for medical image classification.
It exploits the unlabeled data by encouraging the prediction consistency of given input under perturbations.
Our method outperforms many state-of-the-art semi-supervised learning methods on both single-label and multi-label image classification scenarios.
arXiv Detail & Related papers (2020-05-15T06:57:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.