Studying the Effects of Sex-related Differences on Brain Age Prediction
using brain MR Imaging
- URL: http://arxiv.org/abs/2310.11577v1
- Date: Tue, 17 Oct 2023 20:55:53 GMT
- Title: Studying the Effects of Sex-related Differences on Brain Age Prediction
using brain MR Imaging
- Authors: Mahsa Dibaji, Neha Gianchandani, Akhil Nair, Mansi Singhal, Roberto
Souza, Mariana Bento
- Abstract summary: We study biases related to sex when developing a machine learning model based on brain magnetic resonance images (MRI)
We investigate the effects of sex by performing brain age prediction considering different experimental designs.
We found disparities in the performance of brain age prediction models when trained on distinct sex subgroups and datasets.
- Score: 0.3958317527488534
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While utilizing machine learning models, one of the most crucial aspects is
how bias and fairness affect model outcomes for diverse demographics. This
becomes especially relevant in the context of machine learning for medical
imaging applications as these models are increasingly being used for diagnosis
and treatment planning. In this paper, we study biases related to sex when
developing a machine learning model based on brain magnetic resonance images
(MRI). We investigate the effects of sex by performing brain age prediction
considering different experimental designs: model trained using only female
subjects, only male subjects and a balanced dataset. We also perform evaluation
on multiple MRI datasets (Calgary-Campinas(CC359) and CamCAN) to assess the
generalization capability of the proposed models. We found disparities in the
performance of brain age prediction models when trained on distinct sex
subgroups and datasets, in both final predictions and decision making (assessed
using interpretability models). Our results demonstrated variations in model
generalizability across sex-specific subgroups, suggesting potential biases in
models trained on unbalanced datasets. This underlines the critical role of
careful experimental design in generating fair and reliable outcomes.
Related papers
- Dataset Distribution Impacts Model Fairness: Single vs. Multi-Task Learning [2.9530211066840417]
We evaluate the performance of skin lesion classification using ResNet-based CNNs.
We present a linear programming method for generating datasets with varying patient sex and class labels.
arXiv Detail & Related papers (2024-07-24T15:23:26Z) - A Demographic-Conditioned Variational Autoencoder for fMRI Distribution Sampling and Removal of Confounds [49.34500499203579]
We create a variational autoencoder (VAE)-based model, DemoVAE, to decorrelate fMRI features from demographics.
We generate high-quality synthetic fMRI data based on user-supplied demographics.
arXiv Detail & Related papers (2024-05-13T17:49:20Z) - Analysing race and sex bias in brain age prediction [18.68533487971233]
We analyse the commonly used ResNet-34 model by conducting a subgroup performance analysis and feature inspection.
Our results reveal statistically significant differences in predictive performance between Black and White, Black and Asian, and male and female subjects.
arXiv Detail & Related papers (2023-09-19T14:40:19Z) - An investigation into the impact of deep learning model choice on sex
and race bias in cardiac MR segmentation [8.449342469976758]
We investigate how imbalances in subject sex and race affect AI-based cine cardiac magnetic resonance image segmentation.
We find significant sex bias in three of the four models and racial bias in all of the models.
arXiv Detail & Related papers (2023-08-25T14:55:38Z) - Machine Learning Models Are Not Necessarily Biased When Constructed
Properly: Evidence from Neuroimaging Studies [19.288217559980545]
We provide experimental data which support that when properly trained, machine learning models can generalize well across diverse conditions.
Specifically, by using multi-study magnetic resonance imaging consortia for diagnosing Alzheimer's disease, schizophrenia, and autism spectrum disorder, we find that, the accuracy of well-trained models is consistent across different subgroups.
arXiv Detail & Related papers (2022-05-26T15:24:39Z) - Analyzing the Effects of Handling Data Imbalance on Learned Features
from Medical Images by Looking Into the Models [50.537859423741644]
Training a model on an imbalanced dataset can introduce unique challenges to the learning problem.
We look deeper into the internal units of neural networks to observe how handling data imbalance affects the learned features.
arXiv Detail & Related papers (2022-04-04T09:38:38Z) - What Do You See in this Patient? Behavioral Testing of Clinical NLP
Models [69.09570726777817]
We introduce an extendable testing framework that evaluates the behavior of clinical outcome models regarding changes of the input.
We show that model behavior varies drastically even when fine-tuned on the same data and that allegedly best-performing models have not always learned the most medically plausible patterns.
arXiv Detail & Related papers (2021-11-30T15:52:04Z) - Unravelling the Effect of Image Distortions for Biased Prediction of
Pre-trained Face Recognition Models [86.79402670904338]
We evaluate the performance of four state-of-the-art deep face recognition models in the presence of image distortions.
We have observed that image distortions have a relationship with the performance gap of the model across different subgroups.
arXiv Detail & Related papers (2021-08-14T16:49:05Z) - Modeling Shared Responses in Neuroimaging Studies through MultiView ICA [94.31804763196116]
Group studies involving large cohorts of subjects are important to draw general conclusions about brain functional organization.
We propose a novel MultiView Independent Component Analysis model for group studies, where data from each subject are modeled as a linear combination of shared independent sources plus noise.
We demonstrate the usefulness of our approach first on fMRI data, where our model demonstrates improved sensitivity in identifying common sources among subjects.
arXiv Detail & Related papers (2020-06-11T17:29:53Z) - Do Neural Ranking Models Intensify Gender Bias? [13.37092521347171]
We first provide a bias measurement framework which includes two metrics to quantify the degree of the unbalanced presence of gender-related concepts in a given IR model's ranking list.
Applying these queries to the MS MARCO Passage retrieval collection, we then measure the gender bias of a BM25 model and several recent neural ranking models.
Results show that while all models are strongly biased toward male, the neural models, and in particular the ones based on contextualized embedding models, significantly intensify gender bias.
arXiv Detail & Related papers (2020-05-01T13:31:11Z) - InsideBias: Measuring Bias in Deep Networks and Application to Face
Gender Biometrics [73.85525896663371]
This work explores the biases in learning processes based on deep neural network architectures.
We employ two gender detection models based on popular deep neural networks.
We propose InsideBias, a novel method to detect biased models.
arXiv Detail & Related papers (2020-04-14T15:20:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.