An XAI View on Explainable ASP: Methods, Systems, and Perspectives
- URL: http://arxiv.org/abs/2601.14764v1
- Date: Wed, 21 Jan 2026 08:37:33 GMT
- Title: An XAI View on Explainable ASP: Methods, Systems, and Perspectives
- Authors: Thomas Eiter, Tobias Geibinger, Zeynep G. Saribatur,
- Abstract summary: Answer Set Programming (ASP) is a popular declarative reasoning and problem solving approach in symbolic AI.<n>We provide an overview of types of ASP explanations in connection with user questions for explanation, and describe how their coverage by current theory and tools.<n>We pinpoint gaps in existing ASP explanations approaches and identify research directions for future work.
- Score: 16.631602844999723
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Answer Set Programming (ASP) is a popular declarative reasoning and problem solving approach in symbolic AI. Its rule-based formalism makes it inherently attractive for explainable and interpretive reasoning, which is gaining importance with the surge of Explainable AI (XAI). A number of explanation approaches and tools for ASP have been developed, which often tackle specific explanatory settings and may not cover all scenarios that ASP users encounter. In this survey, we provide, guided by an XAI perspective, an overview of types of ASP explanations in connection with user questions for explanation, and describe how their coverage by current theory and tools. Furthermore, we pinpoint gaps in existing ASP explanations approaches and identify research directions for future work.
Related papers
- A Comprehensive Guide to Explainable AI: From Classical Models to LLMs [25.07463077055411]
Explainable Artificial Intelligence (XAI) addresses the growing need for transparency and interpretability in AI systems.<n>It explores interpretability in traditional models like Decision Trees, Linear Regression, and Support Vector Machines.<n>The book presents practical techniques such as SHAP, LIME, Grad-CAM, counterfactual explanations, and causal inference.
arXiv Detail & Related papers (2024-12-01T13:01:01Z) - An Ontology-Enabled Approach For User-Centered and Knowledge-Enabled Explanations of AI Systems [0.3480973072524161]
Recent research in explainability has focused on explaining the workings of AI models or model explainability.
This thesis seeks to bridge some gaps between model and user-centered explainability.
arXiv Detail & Related papers (2024-10-23T02:03:49Z) - Opening the Black-Box: A Systematic Review on Explainable AI in Remote Sensing [51.524108608250074]
Black-box machine learning approaches have become a dominant modeling paradigm for knowledge extraction in remote sensing.
We perform a systematic review to identify the key trends in the field and shed light on novel explainable AI approaches.
We also give a detailed outlook on the challenges and promising research directions.
arXiv Detail & Related papers (2024-02-21T13:19:58Z) - Explainable Answer-set Programming [0.0]
Project aims to fill some of these gaps and contribute to the state of the art in explainable ASP.
We tackle this by extending the language support of existing approaches but also by the development of novel explanation formalisms.
arXiv Detail & Related papers (2023-08-30T09:09:57Z) - Explaining Explainability: Towards Deeper Actionable Insights into Deep
Learning through Second-order Explainability [70.60433013657693]
Second-order explainable AI (SOXAI) was recently proposed to extend explainable AI (XAI) from the instance level to the dataset level.
We demonstrate for the first time, via example classification and segmentation cases, that eliminating irrelevant concepts from the training set based on actionable insights from SOXAI can enhance a model's performance.
arXiv Detail & Related papers (2023-06-14T23:24:01Z) - Towards Human Cognition Level-based Experiment Design for Counterfactual
Explanations (XAI) [68.8204255655161]
The emphasis of XAI research appears to have turned to a more pragmatic explanation approach for better understanding.
An extensive area where cognitive science research may substantially influence XAI advancements is evaluating user knowledge and feedback.
We propose a framework to experiment with generating and evaluating the explanations on the grounds of different cognitive levels of understanding.
arXiv Detail & Related papers (2022-10-31T19:20:22Z) - Tools and Methodologies for Verifying Answer Set Programs [0.0]
ASP is a powerful declarative programming paradigm commonly used for solving challenging search and optimization problems.
As an approach to Knowledge Representation and Reasoning, ASP benefits from its simplicity, conciseness and rigorously defined semantics.
My research is concerned with extending the theory and tools supporting the verification of ASP progams.
arXiv Detail & Related papers (2022-08-05T10:50:21Z) - Making Things Explainable vs Explaining: Requirements and Challenges
under the GDPR [2.578242050187029]
ExplanatorY AI (YAI) builds over XAI with the goal to collect and organize explainable information.
We represent the problem of generating explanations for Automated Decision-Making systems (ADMs) into the identification of an appropriate path over an explanatory space.
arXiv Detail & Related papers (2021-10-02T08:48:47Z) - The Who in XAI: How AI Background Shapes Perceptions of AI Explanations [61.49776160925216]
We conduct a mixed-methods study of how two different groups--people with and without AI background--perceive different types of AI explanations.
We find that (1) both groups showed unwarranted faith in numbers for different reasons and (2) each group found value in different explanations beyond their intended design.
arXiv Detail & Related papers (2021-07-28T17:32:04Z) - Explainability in Deep Reinforcement Learning [68.8204255655161]
We review recent works in the direction to attain Explainable Reinforcement Learning (XRL)
In critical situations where it is essential to justify and explain the agent's behaviour, better explainability and interpretability of RL models could help gain scientific insight on the inner workings of what is still considered a black box.
arXiv Detail & Related papers (2020-08-15T10:11:42Z) - A general framework for scientifically inspired explanations in AI [76.48625630211943]
We instantiate the concept of structure of scientific explanation as the theoretical underpinning for a general framework in which explanations for AI systems can be implemented.
This framework aims to provide the tools to build a "mental-model" of any AI system so that the interaction with the user can provide information on demand and be closer to the nature of human-made explanations.
arXiv Detail & Related papers (2020-03-02T10:32:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.