Harvard Glaucoma Fairness: A Retinal Nerve Disease Dataset for Fairness
Learning and Fair Identity Normalization
- URL: http://arxiv.org/abs/2306.09264v3
- Date: Mon, 11 Mar 2024 02:54:57 GMT
- Title: Harvard Glaucoma Fairness: A Retinal Nerve Disease Dataset for Fairness
Learning and Fair Identity Normalization
- Authors: Yan Luo, Yu Tian, Min Shi, Louis R. Pasquale, Lucy Q. Shen, Nazlee
Zebardast, Tobias Elze, Mengyu Wang
- Abstract summary: We introduce Harvard Glaucoma Fairness (Harvard-GF), a dataset with both 2D and 3D data imaging and balanced racial groups for glaucoma detection.
Our FIN approach is compared with various the-state-of-the-art fairness learning methods with superior performance in the racial, gender, and fairness tasks.
We propose an equity-scaled performance measure, which can be flexibly used to compare all kinds of performance metrics in the context of fairness.
- Score: 13.792327874980632
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Fairness (also known as equity interchangeably) in machine learning is
important for societal well-being, but limited public datasets hinder its
progress. Currently, no dedicated public medical datasets with imaging data for
fairness learning are available, though minority groups suffer from more health
issues. To address this gap, we introduce Harvard Glaucoma Fairness
(Harvard-GF), a retinal nerve disease dataset with both 2D and 3D imaging data
and balanced racial groups for glaucoma detection. Glaucoma is the leading
cause of irreversible blindness globally with Blacks having doubled glaucoma
prevalence than other races. We also propose a fair identity normalization
(FIN) approach to equalize the feature importance between different identity
groups. Our FIN approach is compared with various the-state-of-the-art fairness
learning methods with superior performance in the racial, gender, and ethnicity
fairness tasks with 2D and 3D imaging data, which demonstrate the utilities of
our dataset Harvard-GF for fairness learning. To facilitate fairness
comparisons between different models, we propose an equity-scaled performance
measure, which can be flexibly used to compare all kinds of performance metrics
in the context of fairness. The dataset and code are publicly accessible via
\url{https://ophai.hms.harvard.edu/datasets/harvard-glaucoma-fairness-3300-samples/}.
Related papers
- Fairness without Demographics through Learning Graph of Gradients [22.260763111752805]
We show that the correlation between gradients and groups can help identify and improve group fairness.
Our method is robust to noise and can improve fairness significantly without decreasing the overall accuracy too much.
arXiv Detail & Related papers (2024-12-04T20:35:50Z) - FairCLIP: Harnessing Fairness in Vision-Language Learning [20.743027598445796]
We introduce the first fair vision-language medical dataset that provides detailed demographic attributes, groundtruth labels, and clinical notes.
As the first fair vision-language medical dataset of its kind, HarvardFairMed holds the potential to catalyze the development of machine learning models that are both aware and clinically effective.
arXiv Detail & Related papers (2024-03-29T03:15:31Z) - FairSeg: A Large-Scale Medical Image Segmentation Dataset for Fairness Learning Using Segment Anything Model with Fair Error-Bound Scaling [14.483954095650887]
High-quality medical fairness datasets are needed to promote fairness learning research.
Existing medical fairness datasets are all for classification tasks, and no fairness datasets are available for medical segmentation.
We propose the first fairness dataset for medical segmentation named HarvardFairSeg with 10,000 subject samples.
arXiv Detail & Related papers (2023-11-03T18:44:21Z) - FairVision: Equitable Deep Learning for Eye Disease Screening via Fair Identity Scaling [19.16603153814857]
We conduct the first comprehensive study on the fairness of 3D medical imaging models across multiple protected attributes.
Our investigation spans both 2D and 3D models and evaluates fairness across five architectures on three common eye diseases.
We release Harvard-FairVision, the first large-scale medical fairness dataset with 30,000 subjects featuring both 2D and 3D imaging data.
arXiv Detail & Related papers (2023-10-03T23:44:35Z) - Fairness meets Cross-Domain Learning: a new perspective on Models and
Metrics [80.07271410743806]
We study the relationship between cross-domain learning (CD) and model fairness.
We introduce a benchmark on face and medical images spanning several demographic groups as well as classification and localization tasks.
Our study covers 14 CD approaches alongside three state-of-the-art fairness algorithms and shows how the former can outperform the latter.
arXiv Detail & Related papers (2023-03-25T09:34:05Z) - DualFair: Fair Representation Learning at Both Group and Individual
Levels via Contrastive Self-supervision [73.80009454050858]
This work presents a self-supervised model, called DualFair, that can debias sensitive attributes like gender and race from learned representations.
Our model jointly optimize for two fairness criteria - group fairness and counterfactual fairness.
arXiv Detail & Related papers (2023-03-15T07:13:54Z) - Fairness-aware Model-agnostic Positive and Unlabeled Learning [38.50536380390474]
We propose a fairness-aware Positive and Unlabeled Learning (PUL) method named FairPUL.
For binary classification over individuals from two populations, we aim to achieve similar true positive rates and false positive rates.
Our framework is proven to be statistically consistent in terms of both the classification error and the fairness metric.
arXiv Detail & Related papers (2022-06-19T08:04:23Z) - Fair Group-Shared Representations with Normalizing Flows [68.29997072804537]
We develop a fair representation learning algorithm which is able to map individuals belonging to different groups in a single group.
We show experimentally that our methodology is competitive with other fair representation learning algorithms.
arXiv Detail & Related papers (2022-01-17T10:49:49Z) - MultiFair: Multi-Group Fairness in Machine Learning [52.24956510371455]
We study multi-group fairness in machine learning (MultiFair)
We propose a generic end-to-end algorithmic framework to solve it.
Our proposed framework is generalizable to many different settings.
arXiv Detail & Related papers (2021-05-24T02:30:22Z) - Fairness in Semi-supervised Learning: Unlabeled Data Help to Reduce
Discrimination [53.3082498402884]
A growing specter in the rise of machine learning is whether the decisions made by machine learning models are fair.
We present a framework of fair semi-supervised learning in the pre-processing phase, including pseudo labeling to predict labels for unlabeled data.
A theoretical decomposition analysis of bias, variance and noise highlights the different sources of discrimination and the impact they have on fairness in semi-supervised learning.
arXiv Detail & Related papers (2020-09-25T05:48:56Z) - Deep F-measure Maximization for End-to-End Speech Understanding [52.36496114728355]
We propose a differentiable approximation to the F-measure and train the network with this objective using standard backpropagation.
We perform experiments on two standard fairness datasets, Adult, Communities and Crime, and also on speech-to-intent detection on the ATIS dataset and speech-to-image concept classification on the Speech-COCO dataset.
In all four of these tasks, F-measure results in improved micro-F1 scores, with absolute improvements of up to 8% absolute, as compared to models trained with the cross-entropy loss function.
arXiv Detail & Related papers (2020-08-08T03:02:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.