In Pursuit of Interpretable, Fair and Accurate Machine Learning for
Criminal Recidivism Prediction
- URL: http://arxiv.org/abs/2005.04176v3
- Date: Fri, 11 Mar 2022 23:00:44 GMT
- Title: In Pursuit of Interpretable, Fair and Accurate Machine Learning for
Criminal Recidivism Prediction
- Authors: Caroline Wang, Bin Han, Bhrij Patel, Cynthia Rudin
- Abstract summary: This study trains interpretable models that output probabilities rather than binary predictions, and uses quantitative fairness definitions to assess the models.
We generated black-box and interpretable ML models on two different criminal recidivism datasets from Florida and Kentucky.
Several interpretable ML models can predict recidivism as well as black-box ML models and are more accurate than COMPAS or the Arnold PSA.
- Score: 19.346391120556884
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Objectives: We study interpretable recidivism prediction using machine
learning (ML) models and analyze performance in terms of prediction ability,
sparsity, and fairness. Unlike previous works, this study trains interpretable
models that output probabilities rather than binary predictions, and uses
quantitative fairness definitions to assess the models. This study also
examines whether models can generalize across geographic locations. Methods: We
generated black-box and interpretable ML models on two different criminal
recidivism datasets from Florida and Kentucky. We compared predictive
performance and fairness of these models against two methods that are currently
used in the justice system to predict pretrial recidivism: the Arnold PSA and
COMPAS. We evaluated predictive performance of all models on predicting six
different types of crime over two time spans. Results: Several interpretable ML
models can predict recidivism as well as black-box ML models and are more
accurate than COMPAS or the Arnold PSA. These models are potentially useful in
practice. Similar to the Arnold PSA, some of these interpretable models can be
written down as a simple table. Others can be displayed using a set of
visualizations. Our geographic analysis indicates that ML models should be
trained separately for separate locations and updated over time. We also
present a fairness analysis for the interpretable models. Conclusions:
Interpretable machine learning models can perform just as well as
non-interpretable methods and currently-used risk assessment scales, in terms
of both prediction accuracy and fairness. Machine learning models might be more
accurate when trained separately for distinct locations and kept up-to-date.
Related papers
- Evaluating Model Bias Requires Characterizing its Mistakes [19.777130236160712]
We introduce SkewSize, a principled and flexible metric that captures bias from mistakes in a model's predictions.
It can be used in multi-class settings or generalised to the open vocabulary setting of generative models.
We demonstrate the utility of SkewSize in multiple settings including: standard vision models trained on synthetic data, vision models trained on ImageNet, and large scale vision-and-language models from the BLIP-2 family.
arXiv Detail & Related papers (2024-07-15T11:46:21Z) - Causal Estimation of Memorisation Profiles [58.20086589761273]
Understanding memorisation in language models has practical and societal implications.
Memorisation is the causal effect of training with an instance on the model's ability to predict that instance.
This paper proposes a new, principled, and efficient method to estimate memorisation based on the difference-in-differences design from econometrics.
arXiv Detail & Related papers (2024-06-06T17:59:09Z) - Predictive Churn with the Set of Good Models [64.05949860750235]
We study the effect of conflicting predictions over the set of near-optimal machine learning models.
We present theoretical results on the expected churn between models within the Rashomon set.
We show how our approach can be used to better anticipate, reduce, and avoid churn in consumer-facing applications.
arXiv Detail & Related papers (2024-02-12T16:15:25Z) - EAMDrift: An interpretable self retrain model for time series [0.0]
We present EAMDrift, a novel method that combines forecasts from multiple individual predictors by weighting each prediction according to a performance metric.
EAMDrift is designed to automatically adapt to out-of-distribution patterns in data and identify the most appropriate models to use at each moment.
Our study on real-world datasets shows that EAMDrift outperforms individual baseline models by 20% and achieves comparable accuracy results to non-interpretable ensemble models.
arXiv Detail & Related papers (2023-05-31T13:25:26Z) - Pathologies of Pre-trained Language Models in Few-shot Fine-tuning [50.3686606679048]
We show that pre-trained language models with few examples show strong prediction bias across labels.
Although few-shot fine-tuning can mitigate the prediction bias, our analysis shows models gain performance improvement by capturing non-task-related features.
These observations alert that pursuing model performance with fewer examples may incur pathological prediction behavior.
arXiv Detail & Related papers (2022-04-17T15:55:18Z) - Explain, Edit, and Understand: Rethinking User Study Design for
Evaluating Model Explanations [97.91630330328815]
We conduct a crowdsourcing study, where participants interact with deception detection models that have been trained to distinguish between genuine and fake hotel reviews.
We observe that for a linear bag-of-words model, participants with access to the feature coefficients during training are able to cause a larger reduction in model confidence in the testing phase when compared to the no-explanation control.
arXiv Detail & Related papers (2021-12-17T18:29:56Z) - Why do classifier accuracies show linear trends under distribution
shift? [58.40438263312526]
accuracies of models on one data distribution are approximately linear functions of the accuracies on another distribution.
We assume the probability that two models agree in their predictions is higher than what we can infer from their accuracy levels alone.
We show that a linear trend must occur when evaluating models on two distributions unless the size of the distribution shift is large.
arXiv Detail & Related papers (2020-12-31T07:24:30Z) - Deducing neighborhoods of classes from a fitted model [68.8204255655161]
In this article a new kind of interpretable machine learning method is presented.
It can help to understand the partitioning of the feature space into predicted classes in a classification model using quantile shifts.
Basically, real data points (or specific points of interest) are used and the changes of the prediction after slightly raising or decreasing specific features are observed.
arXiv Detail & Related papers (2020-09-11T16:35:53Z) - A Causal Lens for Peeking into Black Box Predictive Models: Predictive
Model Interpretation via Causal Attribution [3.3758186776249928]
We aim to address this problem in settings where the predictive model is a black box.
We reduce the problem of interpreting a black box predictive model to that of estimating the causal effects of each of the model inputs on the model output.
We show how the resulting causal attribution of responsibility for model output to the different model inputs can be used to interpret the predictive model and to explain its predictions.
arXiv Detail & Related papers (2020-08-01T23:20:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.