Explainability Via Causal Self-Talk
- URL: http://arxiv.org/abs/2211.09937v1
- Date: Thu, 17 Nov 2022 23:17:01 GMT
- Title: Explainability Via Causal Self-Talk
- Authors: Nicholas A. Roy, Junkyung Kim, Neil Rabinowitz
- Abstract summary: Explaining the behavior of AI systems is an important problem that, in practice, is generally avoided.
We describe an effective way to satisfy all the desiderata: train the AI system to build a causal model of itself.
We implement this method in a simulated 3D environment, and show how it enables agents to generate faithful and semantically-meaningful explanations.
- Score: 9.149689942389923
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Explaining the behavior of AI systems is an important problem that, in
practice, is generally avoided. While the XAI community has been developing an
abundance of techniques, most incur a set of costs that the wider deep learning
community has been unwilling to pay in most situations. We take a pragmatic
view of the issue, and define a set of desiderata that capture both the
ambitions of XAI and the practical constraints of deep learning. We describe an
effective way to satisfy all the desiderata: train the AI system to build a
causal model of itself. We develop an instance of this solution for Deep RL
agents: Causal Self-Talk. CST operates by training the agent to communicate
with itself across time. We implement this method in a simulated 3D
environment, and show how it enables agents to generate faithful and
semantically-meaningful explanations of their own behavior. Beyond
explanations, we also demonstrate that these learned models provide new ways of
building semantic control interfaces to AI systems.
Related papers
- Explaining Explaining [0.882727051273924]
Explanation is key to people having confidence in high-stakes AI systems.
Machine-learning-based systems can't explain because they are usually black boxes.
We describe a hybrid approach to developing cognitive agents.
arXiv Detail & Related papers (2024-09-26T16:55:44Z) - Pangu-Agent: A Fine-Tunable Generalist Agent with Structured Reasoning [50.47568731994238]
Key method for creating Artificial Intelligence (AI) agents is Reinforcement Learning (RL)
This paper presents a general framework model for integrating and learning structured reasoning into AI agents' policies.
arXiv Detail & Related papers (2023-12-22T17:57:57Z) - Enabling High-Level Machine Reasoning with Cognitive Neuro-Symbolic
Systems [67.01132165581667]
We propose to enable high-level reasoning in AI systems by integrating cognitive architectures with external neuro-symbolic components.
We illustrate a hybrid framework centered on ACT-R and we discuss the role of generative models in recent and future applications.
arXiv Detail & Related papers (2023-11-13T21:20:17Z) - Towards Reconciling Usability and Usefulness of Explainable AI
Methodologies [2.715884199292287]
Black-box AI systems can lead to liability and accountability issues when they produce an incorrect decision.
Explainable AI (XAI) seeks to bridge the knowledge gap, between developers and end-users.
arXiv Detail & Related papers (2023-01-13T01:08:49Z) - Seamful XAI: Operationalizing Seamful Design in Explainable AI [59.89011292395202]
Mistakes in AI systems are inevitable, arising from both technical limitations and sociotechnical gaps.
We propose that seamful design can foster AI explainability by revealing sociotechnical and infrastructural mismatches.
We explore this process with 43 AI practitioners and real end-users.
arXiv Detail & Related papers (2022-11-12T21:54:05Z) - Responsibility: An Example-based Explainable AI approach via Training
Process Inspection [1.4610038284393165]
We present a novel XAI approach that identifies the most responsible training example for a particular decision.
This example can then be shown as an explanation: "this is what I (the AI) learned that led me to do that"
Our results demonstrate that responsibility can help improve accuracy for both human end users and secondary ML models.
arXiv Detail & Related papers (2022-09-07T19:30:01Z) - A User-Centred Framework for Explainable Artificial Intelligence in
Human-Robot Interaction [70.11080854486953]
We propose a user-centred framework for XAI that focuses on its social-interactive aspect.
The framework aims to provide a structure for interactive XAI solutions thought for non-expert users.
arXiv Detail & Related papers (2021-09-27T09:56:23Z) - Brittle AI, Causal Confusion, and Bad Mental Models: Challenges and
Successes in the XAI Program [17.52385105997044]
Deep neural network driven models have surpassed human level performance in benchmark autonomy tasks.
The underlying policies for these agents, however, are not easily interpretable.
This paper discusses the origins of these takeaways, provides amplifying information, and suggestions for future work.
arXiv Detail & Related papers (2021-06-10T05:21:10Z) - A general framework for scientifically inspired explanations in AI [76.48625630211943]
We instantiate the concept of structure of scientific explanation as the theoretical underpinning for a general framework in which explanations for AI systems can be implemented.
This framework aims to provide the tools to build a "mental-model" of any AI system so that the interaction with the user can provide information on demand and be closer to the nature of human-made explanations.
arXiv Detail & Related papers (2020-03-02T10:32:21Z) - Explainable Active Learning (XAL): An Empirical Study of How Local
Explanations Impact Annotator Experience [76.9910678786031]
We propose a novel paradigm of explainable active learning (XAL), by introducing techniques from the recently surging field of explainable AI (XAI) into an Active Learning setting.
Our study shows benefits of AI explanation as interfaces for machine teaching--supporting trust calibration and enabling rich forms of teaching feedback, and potential drawbacks--anchoring effect with the model judgment and cognitive workload.
arXiv Detail & Related papers (2020-01-24T22:52:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.