A Study of Demographic Bias in CNN-based Brain MR Segmentation
- URL: http://arxiv.org/abs/2208.06613v1
- Date: Sat, 13 Aug 2022 10:07:54 GMT
- Title: A Study of Demographic Bias in CNN-based Brain MR Segmentation
- Authors: Stefanos Ioannou (1), Hana Chockler (1 and 3), Alexander Hammers (2)
and Andrew P. King (2) ((1) Department of Informatics, King's College London,
U.K., (2) School of Biomedical Engineering and Imaging Sciences, King's
College London, U.K., (3) causaLens Ltd., U.K.)
- Abstract summary: CNN models for brain MR segmentation have the potential to contain sex or race bias when trained with imbalanced training sets.
We train multiple instances of the FastSurferCNN model using different levels of sex imbalance in white subjects.
We find significant sex and race bias effects in segmentation model performance.
- Score: 43.55994393060723
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Convolutional neural networks (CNNs) are increasingly being used to automate
the segmentation of brain structures in magnetic resonance (MR) images for
research studies. In other applications, CNN models have been shown to exhibit
bias against certain demographic groups when they are under-represented in the
training sets. In this work, we investigate whether CNN models for brain MR
segmentation have the potential to contain sex or race bias when trained with
imbalanced training sets. We train multiple instances of the FastSurferCNN
model using different levels of sex imbalance in white subjects. We evaluate
the performance of these models separately for white male and white female test
sets to assess sex bias, and furthermore evaluate them on black male and black
female test sets to assess potential racial bias. We find significant sex and
race bias effects in segmentation model performance. The biases have a strong
spatial component, with some brain regions exhibiting much stronger bias than
others. Overall, our results suggest that race bias is more significant than
sex bias. Our study demonstrates the importance of considering race and sex
balance when forming training sets for CNN-based brain MR segmentation, to
avoid maintaining or even exacerbating existing health inequalities through
biased research study findings.
Related papers
- From Bias to Balance: Detecting Facial Expression Recognition Biases in Large Multimodal Foundation Models [3.1927733045184885]
This study addresses the racial biases in facial expression recognition (FER) systems within Large Multimodal Foundation Models (LMFMs)
Existing research predominantly focuses on traditional FER models (CNNs, RNNs, ViTs) leaving a gap in understanding racial biases in LMFMs.
We benchmark four leading LMFMs: GPT-4o, PaliGemma, Gemini, and CLIP to assess their performance in facial emotion detection across different racial demographics.
arXiv Detail & Related papers (2024-08-27T07:54:01Z) - Dataset Distribution Impacts Model Fairness: Single vs. Multi-Task Learning [2.9530211066840417]
We evaluate the performance of skin lesion classification using ResNet-based CNNs.
We present a linear programming method for generating datasets with varying patient sex and class labels.
arXiv Detail & Related papers (2024-07-24T15:23:26Z) - Fast Model Debias with Machine Unlearning [54.32026474971696]
Deep neural networks might behave in a biased manner in many real-world scenarios.
Existing debiasing methods suffer from high costs in bias labeling or model re-training.
We propose a fast model debiasing framework (FMD) which offers an efficient approach to identify, evaluate and remove biases.
arXiv Detail & Related papers (2023-10-19T08:10:57Z) - Analysing race and sex bias in brain age prediction [18.68533487971233]
We analyse the commonly used ResNet-34 model by conducting a subgroup performance analysis and feature inspection.
Our results reveal statistically significant differences in predictive performance between Black and White, Black and Asian, and male and female subjects.
arXiv Detail & Related papers (2023-09-19T14:40:19Z) - The Impact of Debiasing on the Performance of Language Models in
Downstream Tasks is Underestimated [70.23064111640132]
We compare the impact of debiasing on performance across multiple downstream tasks using a wide-range of benchmark datasets.
Experiments show that the effects of debiasing are consistently emphunderestimated across all tasks.
arXiv Detail & Related papers (2023-09-16T20:25:34Z) - Studying Bias in GANs through the Lens of Race [91.95264864405493]
We study how the performance and evaluation of generative image models are impacted by the racial composition of their training datasets.
Our results show that the racial compositions of generated images successfully preserve that of the training data.
However, we observe that truncation, a technique used to generate higher quality images during inference, exacerbates racial imbalances in the data.
arXiv Detail & Related papers (2022-09-06T22:25:56Z) - A systematic study of race and sex bias in CNN-based cardiac MR
segmentation [6.507372382471608]
We present the first systematic study of the impact of training set imbalance on race and sex bias in CNN-based segmentation.
We focus on segmentation of the structures of the heart from short axis cine cardiac magnetic resonance images, and train multiple CNN segmentation models with different levels of race/sex imbalance.
We find no significant bias in the sex experiment but significant bias in two separate race experiments, highlighting the need to consider adequate representation of different demographic groups in health datasets.
arXiv Detail & Related papers (2022-09-04T14:32:00Z) - Mitigating Gender Bias in Captioning Systems [56.25457065032423]
Most captioning models learn gender bias, leading to high gender prediction errors, especially for women.
We propose a new Guided Attention Image Captioning model (GAIC) which provides self-guidance on visual attention to encourage the model to capture correct gender visual evidence.
arXiv Detail & Related papers (2020-06-15T12:16:19Z) - Do Neural Ranking Models Intensify Gender Bias? [13.37092521347171]
We first provide a bias measurement framework which includes two metrics to quantify the degree of the unbalanced presence of gender-related concepts in a given IR model's ranking list.
Applying these queries to the MS MARCO Passage retrieval collection, we then measure the gender bias of a BM25 model and several recent neural ranking models.
Results show that while all models are strongly biased toward male, the neural models, and in particular the ones based on contextualized embedding models, significantly intensify gender bias.
arXiv Detail & Related papers (2020-05-01T13:31:11Z) - InsideBias: Measuring Bias in Deep Networks and Application to Face
Gender Biometrics [73.85525896663371]
This work explores the biases in learning processes based on deep neural network architectures.
We employ two gender detection models based on popular deep neural networks.
We propose InsideBias, a novel method to detect biased models.
arXiv Detail & Related papers (2020-04-14T15:20:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.