How does this interaction affect me? Interpretable attribution for
feature interactions
- URL: http://arxiv.org/abs/2006.10965v1
- Date: Fri, 19 Jun 2020 05:14:24 GMT
- Title: How does this interaction affect me? Interpretable attribution for
feature interactions
- Authors: Michael Tsang, Sirisha Rambhatla, Yan Liu
- Abstract summary: We propose an interaction attribution and detection framework called Archipelago.
Our experiments on standard annotation labels indicate our approach provides significantly more interpretable explanations than comparable methods.
We also provide accompanying visualizations of our approach that give new insights into deep neural networks.
- Score: 19.979889568380464
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine learning transparency calls for interpretable explanations of how
inputs relate to predictions. Feature attribution is a way to analyze the
impact of features on predictions. Feature interactions are the contextual
dependence between features that jointly impact predictions. There are a number
of methods that extract feature interactions in prediction models; however, the
methods that assign attributions to interactions are either uninterpretable,
model-specific, or non-axiomatic. We propose an interaction attribution and
detection framework called Archipelago which addresses these problems and is
also scalable in real-world settings. Our experiments on standard annotation
labels indicate our approach provides significantly more interpretable
explanations than comparable methods, which is important for analyzing the
impact of interactions on predictions. We also provide accompanying
visualizations of our approach that give new insights into deep neural
networks.
Related papers
Err
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.