Finding Uncommon Ground: A Human-Centered Model for Extrospective Explanations
- URL: http://arxiv.org/abs/2507.21571v1
- Date: Tue, 29 Jul 2025 07:59:54 GMT
- Title: Finding Uncommon Ground: A Human-Centered Model for Extrospective Explanations
- Authors: Laura Spillner, Nima Zargham, Mihai Pomarlan, Robert Porzel, Rainer Malaka,
- Abstract summary: AI agents need to focus on individuals and their preferences as well as the context in which the explanations are given.<n>This paper proposes a personalized approach to explanation, where the agent tailors the information provided to the user based on what is most likely pertinent to them.
- Score: 17.427385114802753
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The need for explanations in AI has, by and large, been driven by the desire to increase the transparency of black-box machine learning models. However, such explanations, which focus on the internal mechanisms that lead to a specific output, are often unsuitable for non-experts. To facilitate a human-centered perspective on AI explanations, agents need to focus on individuals and their preferences as well as the context in which the explanations are given. This paper proposes a personalized approach to explanation, where the agent tailors the information provided to the user based on what is most likely pertinent to them. We propose a model of the agent's worldview that also serves as a personal and dynamic memory of its previous interactions with the same user, based on which the artificial agent can estimate what part of its knowledge is most likely new information to the user.
Related papers
- PersonaBench: Evaluating AI Models on Understanding Personal Information through Accessing (Synthetic) Private User Data [76.21047984886273]
Personalization is critical in AI assistants, particularly in the context of private AI models that work with individual users.<n>Due to the sensitive nature of such data, there are no publicly available datasets that allow us to assess an AI model's ability to understand users.<n>We introduce a synthetic data generation pipeline that creates diverse, realistic user profiles and private documents simulating human activities.
arXiv Detail & Related papers (2025-02-28T00:43:35Z) - Memento No More: Coaching AI Agents to Master Multiple Tasks via Hints Internalization [56.674356045200696]
We propose a novel method to train AI agents to incorporate knowledge and skills for multiple tasks without the need for cumbersome note systems or prior high-quality demonstration data.<n>Our approach employs an iterative process where the agent collects new experiences, receives corrective feedback from humans in the form of hints, and integrates this feedback into its weights.<n>We demonstrate the efficacy of our approach by implementing it in a Llama-3-based agent that, after only a few rounds of feedback, outperforms advanced models GPT-4o and DeepSeek-V3 in tasksets.
arXiv Detail & Related papers (2025-02-03T17:45:46Z) - Relevant Irrelevance: Generating Alterfactual Explanations for Image Classifiers [11.200613814162185]
In this paper, we demonstrate the feasibility of alterfactual explanations for black box image classifiers.
We show for the first time that it is possible to apply this idea to black box models based on neural networks.
arXiv Detail & Related papers (2024-05-08T11:03:22Z) - Tell Me More! Towards Implicit User Intention Understanding of Language
Model Driven Agents [110.25679611755962]
Current language model-driven agents often lack mechanisms for effective user participation, which is crucial given the vagueness commonly found in user instructions.
We introduce Intention-in-Interaction (IN3), a novel benchmark designed to inspect users' implicit intentions through explicit queries.
We empirically train Mistral-Interact, a powerful model that proactively assesses task vagueness, inquires user intentions, and refines them into actionable goals.
arXiv Detail & Related papers (2024-02-14T14:36:30Z) - Evaluating the Utility of Model Explanations for Model Development [54.23538543168767]
We evaluate whether explanations can improve human decision-making in practical scenarios of machine learning model development.
To our surprise, we did not find evidence of significant improvement on tasks when users were provided with any of the saliency maps.
These findings suggest caution regarding the usefulness and potential for misunderstanding in saliency-based explanations.
arXiv Detail & Related papers (2023-12-10T23:13:23Z) - Conceptualizing the Relationship between AI Explanations and User Agency [0.9051087836811617]
We analyze the relationship between agency and explanations through a user-centric lens through case studies and thought experiments.
We find that explanation serves as one of several possible first steps for agency by allowing the user convert forethought to outcome in a more effective manner in future interactions.
arXiv Detail & Related papers (2023-12-05T23:56:05Z) - Explaining Agent Behavior with Large Language Models [7.128139268426959]
We propose an approach to generate natural language explanations for an agent's behavior based only on observations of states and actions.
We show how a compact representation of the agent's behavior can be learned and used to produce plausible explanations.
arXiv Detail & Related papers (2023-09-19T06:13:24Z) - Evaluation of Human-Understandability of Global Model Explanations using
Decision Tree [8.263545324859969]
We generate model explanations that are narrative, patient-specific and global.
We find a strong individual preference for a specific type of explanation.
This guides the design of health informatics systems that are both trustworthy and actionable.
arXiv Detail & Related papers (2023-09-18T16:30:14Z) - Explaining Explainability: Towards Deeper Actionable Insights into Deep
Learning through Second-order Explainability [70.60433013657693]
Second-order explainable AI (SOXAI) was recently proposed to extend explainable AI (XAI) from the instance level to the dataset level.
We demonstrate for the first time, via example classification and segmentation cases, that eliminating irrelevant concepts from the training set based on actionable insights from SOXAI can enhance a model's performance.
arXiv Detail & Related papers (2023-06-14T23:24:01Z) - Not all users are the same: Providing personalized explanations for
sequential decision making problems [25.24098967133101]
This work proposes an end-to-end adaptive explanation generation system.
It begins by learning the different types of users that the agent could interact with.
It is then tasked with identifying the type on the fly and adjust its explanations accordingly.
arXiv Detail & Related papers (2021-06-23T07:46:19Z) - A general framework for scientifically inspired explanations in AI [76.48625630211943]
We instantiate the concept of structure of scientific explanation as the theoretical underpinning for a general framework in which explanations for AI systems can be implemented.
This framework aims to provide the tools to build a "mental-model" of any AI system so that the interaction with the user can provide information on demand and be closer to the nature of human-made explanations.
arXiv Detail & Related papers (2020-03-02T10:32:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.