Nutrition Facts, Drug Facts, and Model Facts: Putting AI Ethics into
Practice in Gun Violence Research
- URL: http://arxiv.org/abs/2402.09286v1
- Date: Wed, 14 Feb 2024 16:19:09 GMT
- Title: Nutrition Facts, Drug Facts, and Model Facts: Putting AI Ethics into
Practice in Gun Violence Research
- Authors: Jessica Zhu, Dr. Michel Cukier, Dr. Joseph Richardson Jr
- Abstract summary: We propose a Model Facts template that is easily extendable and decomposes accuracy and demographics into standardized and minimally complex values.
We apply the Model Facts template on two previously published models, a violence risk identification model and a suicide risk prediction model.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Objective: Firearm injury research necessitates using data from
often-exploited vulnerable populations of Black and Brown Americans. In order
to minimize distrust, this study provides a framework for establishing AI trust
and transparency with the general population. Methods: We propose a Model Facts
template that is easily extendable and decomposes accuracy and demographics
into standardized and minimally complex values. This framework allows general
users to assess the validity and biases of a model without diving into
technical model documentation. Examples: We apply the Model Facts template on
two previously published models, a violence risk identification model and a
suicide risk prediction model. We demonstrate the ease of accessing the
appropriate information when the data is structured appropriately. Discussion:
The Model Facts template is limited in its current form to human based data and
biases. Like nutrition facts, it also will require some educational resources
for users to grasp its full utility. Human computer interaction experiments
should be conducted to ensure that the interaction between user interface and
model interface is as desired. Conclusion: The Model Facts label is the first
framework dedicated to establishing trust with end users and general population
consumers. Implementation of Model Facts into firearm injury research will
provide public health practitioners and those impacted by firearm injury
greater faith in the tools the research provides.
Related papers
- Holmes: Towards Effective and Harmless Model Ownership Verification to Personalized Large Vision Models via Decoupling Common Features [54.63343151319368]
This paper proposes a harmless model ownership verification method for personalized models by decoupling similar common features.<n>In the first stage, we create shadow models that retain common features of the victim model while disrupting dataset-specific features.<n>After that, a meta-classifier is trained to identify stolen models by determining whether suspicious models contain the dataset-specific features of the victim.
arXiv Detail & Related papers (2025-06-24T15:40:11Z) - Just as Humans Need Vaccines, So Do Models: Model Immunization to Combat Falsehoods [4.6697477379475005]
Generative AI models often learn and reproduce false information present in their training corpora.<n>This paper argues that AI models should be fine tuned on small, quarantined sets of explicitly labeled falsehoods as a "vaccine" against misinformation.
arXiv Detail & Related papers (2025-05-23T13:20:23Z) - Learning-based Models for Vulnerability Detection: An Extensive Study [3.1317409221921144]
We extensively and comprehensively investigate two types of state-of-the-art learning-based approaches.
We experimentally demonstrate the priority of sequence-based models and the limited abilities of both graph-based models.
arXiv Detail & Related papers (2024-08-14T13:01:30Z) - Privacy Backdoors: Enhancing Membership Inference through Poisoning Pre-trained Models [112.48136829374741]
In this paper, we unveil a new vulnerability: the privacy backdoor attack.
When a victim fine-tunes a backdoored model, their training data will be leaked at a significantly higher rate than if they had fine-tuned a typical model.
Our findings highlight a critical privacy concern within the machine learning community and call for a reevaluation of safety protocols in the use of open-source pre-trained models.
arXiv Detail & Related papers (2024-04-01T16:50:54Z) - Membership Inference Attacks against Synthetic Data through Overfitting
Detection [84.02632160692995]
We argue for a realistic MIA setting that assumes the attacker has some knowledge of the underlying data distribution.
We propose DOMIAS, a density-based MIA model that aims to infer membership by targeting local overfitting of the generative model.
arXiv Detail & Related papers (2023-02-24T11:27:39Z) - Large Language Models with Controllable Working Memory [64.71038763708161]
Large language models (LLMs) have led to a series of breakthroughs in natural language processing (NLP)
What further sets these models apart is the massive amounts of world knowledge they internalize during pretraining.
How the model's world knowledge interacts with the factual information presented in the context remains under explored.
arXiv Detail & Related papers (2022-11-09T18:58:29Z) - Are Neural Topic Models Broken? [81.15470302729638]
We study the relationship between automated and human evaluation of topic models.
We find that neural topic models fare worse in both respects compared to an established classical method.
arXiv Detail & Related papers (2022-10-28T14:38:50Z) - Synthetic Model Combination: An Instance-wise Approach to Unsupervised
Ensemble Learning [92.89846887298852]
Consider making a prediction over new test data without any opportunity to learn from a training set of labelled data.
Give access to a set of expert models and their predictions alongside some limited information about the dataset used to train them.
arXiv Detail & Related papers (2022-10-11T10:20:31Z) - Revealing Unfair Models by Mining Interpretable Evidence [50.48264727620845]
The popularity of machine learning has increased the risk of unfair models getting deployed in high-stake applications.
In this paper, we tackle the novel task of revealing unfair models by mining interpretable evidence.
Our method finds highly interpretable and solid evidence to effectively reveal the unfairness of trained models.
arXiv Detail & Related papers (2022-07-12T20:03:08Z) - Membership Inference Attacks on Deep Regression Models for Neuroimaging [15.591129844038269]
We show realistic Membership Inference attacks on deep learning models trained for 3D neuroimaging tasks in a centralized as well as decentralized setup.
We correctly identified whether an MRI scan was used in model training with a 60% to over 80% success rate depending on model complexity and security assumptions.
arXiv Detail & Related papers (2021-05-06T17:51:06Z) - Interpretabilit\'e des mod\`eles : \'etat des lieux des m\'ethodes et
application \`a l'assurance [1.6058099298620423]
Data is the raw material of many models today make it possible to increase the quality and performance of digital services.
Models users must ensure that models do not discriminate against and that it is also possible to explain its result.
The widening of the panel of predictive algorithms leads scientists to be vigilant about the use of models.
arXiv Detail & Related papers (2020-07-25T12:18:07Z) - Debiasing Skin Lesion Datasets and Models? Not So Fast [17.668005682385175]
Models learned from data risk learning biases from that same data.
When models learn spurious correlations not found in real-world situations, their deployment for critical tasks, such as medical decisions, can be catastrophic.
We find out that, despite interesting results that point to promising future research, current debiasing methods are not ready to solve the bias issue for skin-lesion models.
arXiv Detail & Related papers (2020-04-23T21:07:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.