Towards Transparency in Dermatology Image Datasets with Skin Tone
Annotations by Experts, Crowds, and an Algorithm
- URL: http://arxiv.org/abs/2207.02942v1
- Date: Wed, 6 Jul 2022 19:50:39 GMT
- Title: Towards Transparency in Dermatology Image Datasets with Skin Tone
Annotations by Experts, Crowds, and an Algorithm
- Authors: Matthew Groh, Caleb Harris, Roxana Daneshjou, Omar Badri, Arash
Koochek
- Abstract summary: Public and private image datasets of dermatological conditions rarely include information on skin color.
As a start towards increasing transparency, AI researchers have appropriated the use of the Fitzpatrick skin type (FST) from a measure of patient photosensitivity to a measure for estimating skin tone.
We show that algorithms based on ITA-FST are not reliable for annotating large-scale image datasets.
- Score: 3.6888633946892044
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: While artificial intelligence (AI) holds promise for supporting healthcare
providers and improving the accuracy of medical diagnoses, a lack of
transparency in the composition of datasets exposes AI models to the
possibility of unintentional and avoidable mistakes. In particular, public and
private image datasets of dermatological conditions rarely include information
on skin color. As a start towards increasing transparency, AI researchers have
appropriated the use of the Fitzpatrick skin type (FST) from a measure of
patient photosensitivity to a measure for estimating skin tone in algorithmic
audits of computer vision applications including facial recognition and
dermatology diagnosis. In order to understand the variability of estimated FST
annotations on images, we compare several FST annotation methods on a diverse
set of 460 images of skin conditions from both textbooks and online dermatology
atlases. We find the inter-rater reliability between three board-certified
dermatologists is comparable to the inter-rater reliability between the
board-certified dermatologists and two crowdsourcing methods. In contrast, we
find that the Individual Typology Angle converted to FST (ITA-FST) method
produces annotations that are significantly less correlated with the experts'
annotations than the experts' annotations are correlated with each other. These
results demonstrate that algorithms based on ITA-FST are not reliable for
annotating large-scale image datasets, but human-centered, crowd-based
protocols can reliably add skin type transparency to dermatology datasets.
Furthermore, we introduce the concept of dynamic consensus protocols with
tunable parameters including expert review that increase the visibility of
crowdwork and provide guidance for future crowdsourced annotations of large
image datasets.
Related papers
- Enhancing Skin Disease Diagnosis: Interpretable Visual Concept Discovery with SAM Empowerment [41.398287899966995]
Current AI-assisted skin image diagnosis has achieved dermatologist-level performance in classifying skin cancer.
We propose a novel Cross-Attentive Fusion framework for interpretable skin lesion diagnosis.
arXiv Detail & Related papers (2024-09-14T20:11:25Z) - HistoSPACE: Histology-Inspired Spatial Transcriptome Prediction And Characterization Engine [0.0]
HistoSPACE model explore the diversity of histological images available with ST data to extract molecular insights from tissue image.
Model demonstrates significant efficiency compared to contemporary algorithms, revealing a correlation of 0.56 in leave-one-out cross-validation.
arXiv Detail & Related papers (2024-08-07T07:12:52Z) - SkinGEN: an Explainable Dermatology Diagnosis-to-Generation Framework with Interactive Vision-Language Models [52.90397538472582]
SkinGEN is a diagnosis-to-generation framework that generates reference demonstrations from diagnosis results provided by VLM.
We conduct a user study with 32 participants evaluating both the system performance and explainability.
Results demonstrate that SkinGEN significantly improves users' comprehension of VLM predictions and fosters increased trust in the diagnostic process.
arXiv Detail & Related papers (2024-04-23T05:36:33Z) - On the notion of Hallucinations from the lens of Bias and Validity in
Synthetic CXR Images [0.35998666903987897]
Generative models, such as diffusion models, aim to mitigate data quality and clinical information disparities.
At Stanford, researchers explored the utility of a fine-tuned Stable Diffusion model (RoentGen) for medical imaging data augmentation.
We leveraged RoentGen to produce synthetic Chest-XRay (CXR) images and conducted assessments on bias, validity, and hallucinations.
arXiv Detail & Related papers (2023-12-12T04:41:20Z) - Robust and Interpretable Medical Image Classifiers via Concept
Bottleneck Models [49.95603725998561]
We propose a new paradigm to build robust and interpretable medical image classifiers with natural language concepts.
Specifically, we first query clinical concepts from GPT-4, then transform latent image features into explicit concepts with a vision-language model.
arXiv Detail & Related papers (2023-10-04T21:57:09Z) - High Fidelity Synthetic Face Generation for Rosacea Skin Condition from
Limited Data [0.27998963147546135]
StyleGANs, mainly variants of StyleGANs, have demonstrated promising results in generating synthetic facial images.
A small dataset of Rosacea with 300 full-face images is utilized to further investigate the possibility of generating synthetic data.
The preliminary experiments show how fine-tuning the model and varying experimental settings significantly affect the fidelity of the Rosacea features.
arXiv Detail & Related papers (2023-03-08T19:18:58Z) - PCA: Semi-supervised Segmentation with Patch Confidence Adversarial
Training [52.895952593202054]
We propose a new semi-supervised adversarial method called Patch Confidence Adrial Training (PCA) for medical image segmentation.
PCA learns the pixel structure and context information in each patch to get enough gradient feedback, which aids the discriminator in convergent to an optimal state.
Our method outperforms the state-of-the-art semi-supervised methods, which demonstrates its effectiveness for medical image segmentation.
arXiv Detail & Related papers (2022-07-24T07:45:47Z) - Deep Co-Attention Network for Multi-View Subspace Learning [73.3450258002607]
We propose a deep co-attention network for multi-view subspace learning.
It aims to extract both the common information and the complementary information in an adversarial setting.
In particular, it uses a novel cross reconstruction loss and leverages the label information to guide the construction of the latent representation.
arXiv Detail & Related papers (2021-02-15T18:46:44Z) - Proactive Pseudo-Intervention: Causally Informed Contrastive Learning
For Interpretable Vision Models [103.64435911083432]
We present a novel contrastive learning strategy called it Proactive Pseudo-Intervention (PPI)
PPI leverages proactive interventions to guard against image features with no causal relevance.
We also devise a novel causally informed salience mapping module to identify key image pixels to intervene, and show it greatly facilitates model interpretability.
arXiv Detail & Related papers (2020-12-06T20:30:26Z) - Potential Features of ICU Admission in X-ray Images of COVID-19 Patients [8.83608410540057]
This paper presents an original methodology for extracting semantic features that correlate to severity from a data set with patient ICU admission labels.
The methodology employs a neural network trained to recognise lung pathologies to extract the semantic features.
The method has shown to be capable of selecting images for the learned features, which could translate some information about their common locations in the lung.
arXiv Detail & Related papers (2020-09-26T13:48:39Z) - Semi-supervised Medical Image Classification with Relation-driven
Self-ensembling Model [71.80319052891817]
We present a relation-driven semi-supervised framework for medical image classification.
It exploits the unlabeled data by encouraging the prediction consistency of given input under perturbations.
Our method outperforms many state-of-the-art semi-supervised learning methods on both single-label and multi-label image classification scenarios.
arXiv Detail & Related papers (2020-05-15T06:57:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.