Memory networks for consumer protection:unfairness exposed
- URL: http://arxiv.org/abs/2008.07346v1
- Date: Fri, 24 Jul 2020 14:25:54 GMT
- Title: Memory networks for consumer protection:unfairness exposed
- Authors: Federico Ruggeri, Francesca Lagioia, Marco Lippi, Paolo Torroni
- Abstract summary: Data-driven AI methods can leverage consumer protection by supporting the automated analysis of legal documents.
We consider several configurations of memory-augmented neural networks where rationales are given a special role in the modeling of context knowledge.
Our results show that rationales not only contribute to improve the classification accuracy, but are also able to offer meaningful, natural language explanations.
- Score: 12.884439911728112
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent work has demonstrated how data-driven AI methods can leverage consumer
protection by supporting the automated analysis of legal documents. However, a
shortcoming of data-driven approaches is poor explainability. We posit that in
this domain useful explanations of classifier outcomes can be provided by
resorting to legal rationales. We thus consider several configurations of
memory-augmented neural networks where rationales are given a special role in
the modeling of context knowledge. Our results show that rationales not only
contribute to improve the classification accuracy, but are also able to offer
meaningful, natural language explanations of otherwise opaque classifier
outcomes.
Related papers
- LaPLACE: Probabilistic Local Model-Agnostic Causal Explanations [1.0370398945228227]
We introduce LaPLACE-explainer, designed to provide probabilistic cause-and-effect explanations for machine learning models.
The LaPLACE-Explainer component leverages the concept of a Markov blanket to establish statistical boundaries between relevant and non-relevant features.
Our approach offers causal explanations and outperforms LIME and SHAP in terms of local accuracy and consistency of explained features.
arXiv Detail & Related papers (2023-10-01T04:09:59Z) - Explaining Explainability: Towards Deeper Actionable Insights into Deep
Learning through Second-order Explainability [70.60433013657693]
Second-order explainable AI (SOXAI) was recently proposed to extend explainable AI (XAI) from the instance level to the dataset level.
We demonstrate for the first time, via example classification and segmentation cases, that eliminating irrelevant concepts from the training set based on actionable insights from SOXAI can enhance a model's performance.
arXiv Detail & Related papers (2023-06-14T23:24:01Z) - Discriminative Attribution from Counterfactuals [64.94009515033984]
We present a method for neural network interpretability by combining feature attribution with counterfactual explanations.
We show that this method can be used to quantitatively evaluate the performance of feature attribution methods in an objective manner.
arXiv Detail & Related papers (2021-09-28T00:53:34Z) - Explainable Adversarial Attacks in Deep Neural Networks Using Activation
Profiles [69.9674326582747]
This paper presents a visual framework to investigate neural network models subjected to adversarial examples.
We show how observing these elements can quickly pinpoint exploited areas in a model.
arXiv Detail & Related papers (2021-03-18T13:04:21Z) - A Taxonomy of Explainable Bayesian Networks [0.0]
We introduce a taxonomy of explainability in Bayesian networks.
We extend the existing categorisation of explainability in the model, reasoning or evidence to include explanation of decisions.
arXiv Detail & Related papers (2021-01-28T07:29:57Z) - Generative Counterfactuals for Neural Networks via Attribute-Informed
Perturbation [51.29486247405601]
We design a framework to generate counterfactuals for raw data instances with the proposed Attribute-Informed Perturbation (AIP)
By utilizing generative models conditioned with different attributes, counterfactuals with desired labels can be obtained effectively and efficiently.
Experimental results on real-world texts and images demonstrate the effectiveness, sample quality as well as efficiency of our designed framework.
arXiv Detail & Related papers (2021-01-18T08:37:13Z) - Latent-CF: A Simple Baseline for Reverse Counterfactual Explanations [0.0]
High quality explanations are first step in assessing fairness.
It is important to find a baseline for producing them.
We show that latent space counterfactual generation strikes balance between the speed of basic feature gradient descent methods and authenticity of counterfactuals generated by more complex feature space oriented techniques.
arXiv Detail & Related papers (2020-12-16T22:35:42Z) - Interpretable Representations in Explainable AI: From Theory to Practice [7.031336702345381]
Interpretable representations are the backbone of many explainers that target black-box predictive systems.
We study properties of interpretable representations that encode presence and absence of human-comprehensible concepts.
arXiv Detail & Related papers (2020-08-16T21:44:03Z) - Evaluations and Methods for Explanation through Robustness Analysis [117.7235152610957]
We establish a novel set of evaluation criteria for such feature based explanations by analysis.
We obtain new explanations that are loosely necessary and sufficient for a prediction.
We extend the explanation to extract the set of features that would move the current prediction to a target class.
arXiv Detail & Related papers (2020-05-31T05:52:05Z) - Explainable Deep Classification Models for Domain Generalization [94.43131722655617]
Explanations are defined as regions of visual evidence upon which a deep classification network makes a decision.
Our training strategy enforces a periodic saliency-based feedback to encourage the model to focus on the image regions that directly correspond to the ground-truth object.
arXiv Detail & Related papers (2020-03-13T22:22:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.