Learning Models for Actionable Recourse
- URL: http://arxiv.org/abs/2011.06146v3
- Date: Sat, 19 Mar 2022 19:27:57 GMT
- Title: Learning Models for Actionable Recourse
- Authors: Alexis Ross, Himabindu Lakkaraju, Osbert Bastani
- Abstract summary: We propose an algorithm that theoretically guarantees recourse to affected individuals with high probability without sacrificing accuracy.
We demonstrate the efficacy of our approach via extensive experiments on real data.
- Score: 31.30850378503406
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As machine learning models are increasingly deployed in high-stakes domains
such as legal and financial decision-making, there has been growing interest in
post-hoc methods for generating counterfactual explanations. Such explanations
provide individuals adversely impacted by predicted outcomes (e.g., an
applicant denied a loan) with recourse -- i.e., a description of how they can
change their features to obtain a positive outcome. We propose a novel
algorithm that leverages adversarial training and PAC confidence sets to learn
models that theoretically guarantee recourse to affected individuals with high
probability without sacrificing accuracy. We demonstrate the efficacy of our
approach via extensive experiments on real data.
Related papers
- Bank Loan Prediction Using Machine Learning Techniques [0.0]
We have worked on a dataset of 148,670 instances and 37 attributes using machine learning methods.
The best-performing algorithm was AdaBoosting, which achieved an incredible accuracy of 99.99%.
arXiv Detail & Related papers (2024-10-11T15:01:47Z) - Inclusive FinTech Lending via Contrastive Learning and Domain Adaptation [9.75150920742607]
FinTech lending has played a significant role in facilitating financial inclusion.
There are concerns about the potentially biased algorithmic decision-making during loan screening.
We propose a new Transformer-based sequential loan screening model with self-supervised contrastive learning and domain adaptation.
arXiv Detail & Related papers (2023-05-10T01:11:35Z) - Counterfactual Fair Opportunity: Measuring Decision Model Fairness with
Counterfactual Reasoning [5.626570248105078]
This work aims to unveil unfair model behaviors using counterfactual reasoning in the case of fairness under unawareness setting.
A counterfactual version of equal opportunity named counterfactual fair opportunity is defined and two novel metrics that analyze the sensitive information of counterfactual samples are introduced.
arXiv Detail & Related papers (2023-02-16T09:13:53Z) - Robust Transferable Feature Extractors: Learning to Defend Pre-Trained
Networks Against White Box Adversaries [69.53730499849023]
We show that adversarial examples can be successfully transferred to another independently trained model to induce prediction errors.
We propose a deep learning-based pre-processing mechanism, which we refer to as a robust transferable feature extractor (RTFE)
arXiv Detail & Related papers (2022-09-14T21:09:34Z) - Explain, Edit, and Understand: Rethinking User Study Design for
Evaluating Model Explanations [97.91630330328815]
We conduct a crowdsourcing study, where participants interact with deception detection models that have been trained to distinguish between genuine and fake hotel reviews.
We observe that for a linear bag-of-words model, participants with access to the feature coefficients during training are able to cause a larger reduction in model confidence in the testing phase when compared to the no-explanation control.
arXiv Detail & Related papers (2021-12-17T18:29:56Z) - Individual Explanations in Machine Learning Models: A Survey for
Practitioners [69.02688684221265]
The use of sophisticated statistical models that influence decisions in domains of high societal relevance is on the rise.
Many governments, institutions, and companies are reluctant to their adoption as their output is often difficult to explain in human-interpretable ways.
Recently, the academic literature has proposed a substantial amount of methods for providing interpretable explanations to machine learning models.
arXiv Detail & Related papers (2021-04-09T01:46:34Z) - Beyond Trivial Counterfactual Explanations with Diverse Valuable
Explanations [64.85696493596821]
In computer vision applications, generative counterfactual methods indicate how to perturb a model's input to change its prediction.
We propose a counterfactual method that learns a perturbation in a disentangled latent space that is constrained using a diversity-enforcing loss.
Our model improves the success rate of producing high-quality valuable explanations when compared to previous state-of-the-art methods.
arXiv Detail & Related papers (2021-03-18T12:57:34Z) - Accurate and Robust Feature Importance Estimation under Distribution
Shifts [49.58991359544005]
PRoFILE is a novel feature importance estimation method.
We show significant improvements over state-of-the-art approaches, both in terms of fidelity and robustness.
arXiv Detail & Related papers (2020-09-30T05:29:01Z) - Incentives for Federated Learning: a Hypothesis Elicitation Approach [10.452709936265274]
Federated learning provides a promising paradigm for collecting machine learning models from distributed data sources.
The success of a credible federated learning system builds on the assumption that the decentralized and self-interested users will be willing to participate.
This paper introduces solutions to incentivize truthful reporting of a local, user-side machine learning model.
arXiv Detail & Related papers (2020-07-21T04:55:31Z) - Plausible Counterfactuals: Auditing Deep Learning Classifiers with
Realistic Adversarial Examples [84.8370546614042]
Black-box nature of Deep Learning models has posed unanswered questions about what they learn from data.
Generative Adversarial Network (GAN) and multi-objectives are used to furnish a plausible attack to the audited model.
Its utility is showcased within a human face classification task, unveiling the enormous potential of the proposed framework.
arXiv Detail & Related papers (2020-03-25T11:08:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.