An ASP-Based Approach to Counterfactual Explanations for Classification
- URL: http://arxiv.org/abs/2004.13237v2
- Date: Tue, 16 Jun 2020 03:56:13 GMT
- Title: An ASP-Based Approach to Counterfactual Explanations for Classification
- Authors: Leopoldo Bertossi
- Abstract summary: We propose answer-set programs that specify and compute counterfactual interventions as a basis for causality-based explanations to decisions produced by classification models.
They can be applied with black-box models and models that can be specified as logic programs, such as rule-based classifiers.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose answer-set programs that specify and compute counterfactual
interventions as a basis for causality-based explanations to decisions produced
by classification models. They can be applied with black-box models and models
that can be specified as logic programs, such as rule-based classifiers. The
main focus in on the specification and computation of maximum responsibility
causal explanations. The use of additional semantic knowledge is investigated.
Related papers
- The Foundations of Tokenization: Statistical and Computational Concerns [51.370165245628975]
Tokenization is a critical step in the NLP pipeline.
Despite its recognized importance as a standard representation method in NLP, the theoretical underpinnings of tokenization are not yet fully understood.
The present paper contributes to addressing this theoretical gap by proposing a unified formal framework for representing and analyzing tokenizer models.
arXiv Detail & Related papers (2024-07-16T11:12:28Z) - Evaluating Human Alignment and Model Faithfulness of LLM Rationale [66.75309523854476]
We study how well large language models (LLMs) explain their generations through rationales.
We show that prompting-based methods are less "faithful" than attribution-based explanations.
arXiv Detail & Related papers (2024-06-28T20:06:30Z) - Selecting Interpretability Techniques for Healthcare Machine Learning models [69.65384453064829]
In healthcare there is a pursuit for employing interpretable algorithms to assist healthcare professionals in several decision scenarios.
We overview a selection of eight algorithms, both post-hoc and model-based, that can be used for such purposes.
arXiv Detail & Related papers (2024-06-14T17:49:04Z) - Explainability for Large Language Models: A Survey [59.67574757137078]
Large language models (LLMs) have demonstrated impressive capabilities in natural language processing.
This paper introduces a taxonomy of explainability techniques and provides a structured overview of methods for explaining Transformer-based language models.
arXiv Detail & Related papers (2023-09-02T22:14:26Z) - Explainability in Process Outcome Prediction: Guidelines to Obtain
Interpretable and Faithful Models [77.34726150561087]
We define explainability through the interpretability of the explanations and the faithfulness of the explainability model in the field of process outcome prediction.
This paper contributes a set of guidelines named X-MOP which allows selecting the appropriate model based on the event log specifications.
arXiv Detail & Related papers (2022-03-30T05:59:50Z) - Reasoning about Counterfactuals and Explanations: Problems, Results and
Directions [0.0]
These approaches are flexible and modular in that they allow the seamless addition of domain knowledge.
The programs can be used to specify and compute responsibility-based numerical scores as attributive explanations for classification results.
arXiv Detail & Related papers (2021-08-25T01:04:49Z) - Answer-Set Programs for Reasoning about Counterfactual Interventions and
Responsibility Scores for Classification [0.0]
We describe how answer-set programs can be used to declaratively specify counterfactual interventions on entities under classification.
In particular, they can be used to define and compute responsibility scores as attribution-based explanations for outcomes from classification models.
arXiv Detail & Related papers (2021-07-21T15:41:56Z) - Score-Based Explanations in Data Management and Machine Learning: An
Answer-Set Programming Approach to Counterfactual Analysis [0.0]
We describe some recent approaches to score-based explanations for query answers in databases and outcomes from classification models in machine learning.
Special emphasis is placed on declarative approaches based on answer-set programming to the use of counterfactual reasoning for score specification and computation.
arXiv Detail & Related papers (2021-06-19T19:21:48Z) - Declarative Approaches to Counterfactual Explanations for Classification [0.0]
We propose answer-set programs that specify and compute counterfactual interventions on entities that are input on a classification model.
The resulting counterfactual entities serve as a basis for the definition and computation of causality-based explanation scores for the feature values in the entity under classification.
arXiv Detail & Related papers (2020-11-15T00:44:33Z) - Score-Based Explanations in Data Management and Machine Learning [0.0]
We consider explanations for query answers in databases, and for results from classification models.
The described approaches are mostly of a causal and counterfactual nature.
arXiv Detail & Related papers (2020-07-24T23:13:27Z) - Marginal likelihood computation for model selection and hypothesis
testing: an extensive review [66.37504201165159]
This article provides a comprehensive study of the state-of-the-art of the topic.
We highlight limitations, benefits, connections and differences among the different techniques.
Problems and possible solutions with the use of improper priors are also described.
arXiv Detail & Related papers (2020-05-17T18:31:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.