Multi-Objective Counterfactual Explanations
- URL: http://arxiv.org/abs/2004.11165v2
- Date: Wed, 24 Jun 2020 10:03:01 GMT
- Title: Multi-Objective Counterfactual Explanations
- Authors: Susanne Dandl, Christoph Molnar, Martin Binder and Bernd Bischl
- Abstract summary: We propose the Multi-Objective Counterfactuals (MOC) method, which translates the counterfactual search into a multi-objective optimization problem.
Our approach not only returns a diverse set of counterfactuals with different trade-offs between the proposed objectives, but also maintains diversity in feature space.
- Score: 0.7349727826230864
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Counterfactual explanations are one of the most popular methods to make
predictions of black box machine learning models interpretable by providing
explanations in the form of `what-if scenarios'. Most current approaches
optimize a collapsed, weighted sum of multiple objectives, which are naturally
difficult to balance a-priori. We propose the Multi-Objective Counterfactuals
(MOC) method, which translates the counterfactual search into a multi-objective
optimization problem. Our approach not only returns a diverse set of
counterfactuals with different trade-offs between the proposed objectives, but
also maintains diversity in feature space. This enables a more detailed
post-hoc analysis to facilitate better understanding and also more options for
actionable user responses to change the predicted outcome. Our approach is also
model-agnostic and works for numerical and categorical input features. We show
the usefulness of MOC in concrete cases and compare our approach with
state-of-the-art methods for counterfactual explanations.
Related papers
- UCB-driven Utility Function Search for Multi-objective Reinforcement Learning [75.11267478778295]
In Multi-objective Reinforcement Learning (MORL) agents are tasked with optimising decision-making behaviours.
We focus on the case of linear utility functions parameterised by weight vectors w.
We introduce a method based on Upper Confidence Bound to efficiently search for the most promising weight vectors during different stages of the learning process.
arXiv Detail & Related papers (2024-05-01T09:34:42Z) - A multi-criteria approach for selecting an explanation from the set of counterfactuals produced by an ensemble of explainers [4.239829789304117]
We propose to use a multi-stage ensemble approach that will select single counterfactual based on the multiple-criteria analysis.
The proposed approach generates fully actionable counterfactuals with attractive compromise values of the considered quality measures.
arXiv Detail & Related papers (2024-03-20T19:25:11Z) - Faithful Explanations of Black-box NLP Models Using LLM-generated
Counterfactuals [67.64770842323966]
Causal explanations of predictions of NLP systems are essential to ensure safety and establish trust.
Existing methods often fall short of explaining model predictions effectively or efficiently.
We propose two approaches for counterfactual (CF) approximation.
arXiv Detail & Related papers (2023-10-01T07:31:04Z) - MinT: Boosting Generalization in Mathematical Reasoning via Multi-View
Fine-Tuning [53.90744622542961]
Reasoning in mathematical domains remains a significant challenge for small language models (LMs)
We introduce a new method that exploits existing mathematical problem datasets with diverse annotation styles.
Experimental results show that our strategy enables a LLaMA-7B model to outperform prior approaches.
arXiv Detail & Related papers (2023-07-16T05:41:53Z) - An Additive Instance-Wise Approach to Multi-class Model Interpretation [53.87578024052922]
Interpretable machine learning offers insights into what factors drive a certain prediction of a black-box system.
Existing methods mainly focus on selecting explanatory input features, which follow either locally additive or instance-wise approaches.
This work exploits the strengths of both methods and proposes a global framework for learning local explanations simultaneously for multiple target classes.
arXiv Detail & Related papers (2022-07-07T06:50:27Z) - MACE: An Efficient Model-Agnostic Framework for Counterfactual
Explanation [132.77005365032468]
We propose a novel framework of Model-Agnostic Counterfactual Explanation (MACE)
In our MACE approach, we propose a novel RL-based method for finding good counterfactual examples and a gradient-less descent method for improving proximity.
Experiments on public datasets validate the effectiveness with better validity, sparsity and proximity.
arXiv Detail & Related papers (2022-05-31T04:57:06Z) - Model-agnostic and Scalable Counterfactual Explanations via
Reinforcement Learning [0.5729426778193398]
We propose a deep reinforcement learning approach that transforms the optimization procedure into an end-to-end learnable process.
Our experiments on real-world data show that our method is model-agnostic, relying only on feedback from model predictions.
arXiv Detail & Related papers (2021-06-04T16:54:36Z) - Optimal Counterfactual Explanations for Scorecard modelling [0.0]
In this work, we investigate mathematical programming formulations for scorecard models.
The proposed mixed-integer programming formulations combine objective functions to ensure close, realistic and sparse counterfactuals.
Experiments on two real-world datasets confirm that the presented approach can generate optimal diverse counterfactuals.
arXiv Detail & Related papers (2021-04-17T18:51:50Z) - Beyond Trivial Counterfactual Explanations with Diverse Valuable
Explanations [64.85696493596821]
In computer vision applications, generative counterfactual methods indicate how to perturb a model's input to change its prediction.
We propose a counterfactual method that learns a perturbation in a disentangled latent space that is constrained using a diversity-enforcing loss.
Our model improves the success rate of producing high-quality valuable explanations when compared to previous state-of-the-art methods.
arXiv Detail & Related papers (2021-03-18T12:57:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.