Instance-wise or Class-wise? A Tale of Neighbor Shapley for
Concept-based Explanation
- URL: http://arxiv.org/abs/2109.01369v1
- Date: Fri, 3 Sep 2021 08:34:37 GMT
- Title: Instance-wise or Class-wise? A Tale of Neighbor Shapley for
Concept-based Explanation
- Authors: Jiahui Li, Kun Kuang, Lin Li, Long Chen, Songyang Zhang, Jian Shao,
Jun Xiao
- Abstract summary: Deep neural networks have demonstrated remarkable performance in many data-driven and prediction-oriented applications.
Their most significant drawback is the lack of interpretability, which makes them less attractive in many real-world applications.
- Score: 37.033629287045784
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks have demonstrated remarkable performance in many
data-driven and prediction-oriented applications, and sometimes even perform
better than humans. However, their most significant drawback is the lack of
interpretability, which makes them less attractive in many real-world
applications. When relating to the moral problem or the environmental factors
that are uncertain such as crime judgment, financial analysis, and medical
diagnosis, it is essential to mine the evidence for the model's prediction
(interpret model knowledge) to convince humans. Thus, investigating how to
interpret model knowledge is of paramount importance for both academic research
and real applications.
Related papers
- Evaluating the Utility of Model Explanations for Model Development [54.23538543168767]
We evaluate whether explanations can improve human decision-making in practical scenarios of machine learning model development.
To our surprise, we did not find evidence of significant improvement on tasks when users were provided with any of the saliency maps.
These findings suggest caution regarding the usefulness and potential for misunderstanding in saliency-based explanations.
arXiv Detail & Related papers (2023-12-10T23:13:23Z) - Learning for Counterfactual Fairness from Observational Data [62.43249746968616]
Fairness-aware machine learning aims to eliminate biases of learning models against certain subgroups described by certain protected (sensitive) attributes such as race, gender, and age.
A prerequisite for existing methods to achieve counterfactual fairness is the prior human knowledge of the causal model for the data.
In this work, we address the problem of counterfactually fair prediction from observational data without given causal models by proposing a novel framework CLAIRE.
arXiv Detail & Related papers (2023-07-17T04:08:29Z) - A Survey of Deep Causal Models and Their Industrial Applications [5.459987844611099]
This review mainly focuses on the overview of the deep causal models based on neural networks.
We outline some typical applications of causal effect estimation to industry.
arXiv Detail & Related papers (2022-09-19T09:05:05Z) - Causal Reasoning Meets Visual Representation Learning: A Prospective
Study [117.08431221482638]
Lack of interpretability, robustness, and out-of-distribution generalization are becoming the challenges of the existing visual models.
Inspired by the strong inference ability of human-level agents, recent years have witnessed great effort in developing causal reasoning paradigms.
This paper aims to provide a comprehensive overview of this emerging field, attract attention, encourage discussions, bring to the forefront the urgency of developing novel causal reasoning methods.
arXiv Detail & Related papers (2022-04-26T02:22:28Z) - Mixture of Linear Models Co-supervised by Deep Neural Networks [14.831346286039151]
We propose an approach to fill the gap between relatively simple explainable models and deep neural network (DNN) models.
Our main idea is a mixture of discriminative models that is trained with the guidance from a DNN.
arXiv Detail & Related papers (2021-08-05T02:08:35Z) - Causality in Neural Networks -- An Extended Abstract [0.0]
Causal reasoning is the main learning and explanation tool used by humans.
Introducing the ideas of causality to machine learning helps in providing better learning and explainable models.
arXiv Detail & Related papers (2021-06-03T09:52:36Z) - Individual Explanations in Machine Learning Models: A Survey for
Practitioners [69.02688684221265]
The use of sophisticated statistical models that influence decisions in domains of high societal relevance is on the rise.
Many governments, institutions, and companies are reluctant to their adoption as their output is often difficult to explain in human-interpretable ways.
Recently, the academic literature has proposed a substantial amount of methods for providing interpretable explanations to machine learning models.
arXiv Detail & Related papers (2021-04-09T01:46:34Z) - Evaluating the Interpretability of Generative Models by Interactive
Reconstruction [30.441247705313575]
We introduce a task to quantify the human-interpretability of generative model representations.
We find performance on this task much more reliably differentiates entangled and disentangled models than baseline approaches.
arXiv Detail & Related papers (2021-02-02T02:38:14Z) - Plausible Counterfactuals: Auditing Deep Learning Classifiers with
Realistic Adversarial Examples [84.8370546614042]
Black-box nature of Deep Learning models has posed unanswered questions about what they learn from data.
Generative Adversarial Network (GAN) and multi-objectives are used to furnish a plausible attack to the audited model.
Its utility is showcased within a human face classification task, unveiling the enormous potential of the proposed framework.
arXiv Detail & Related papers (2020-03-25T11:08:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.