Tradeoff-Focused Contrastive Explanation for MDP Planning
- URL: http://arxiv.org/abs/2004.12960v2
- Date: Sun, 2 Aug 2020 16:07:02 GMT
- Title: Tradeoff-Focused Contrastive Explanation for MDP Planning
- Authors: Roykrong Sukkerd, Reid Simmons, and David Garlan
- Abstract summary: In real-world applications of planning, planning agents' decisions can involve complex tradeoffs among competing objectives.
It can be difficult for the end-users to understand why an agent decides on a particular planning solution on the basis of its objective values.
We propose an approach, based on contrastive explanation, that enables a multi-objective MDP planning agent to explain its decisions in a way that communicates its tradeoff rationale.
- Score: 7.929642367937801
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: End-users' trust in automated agents is important as automated
decision-making and planning is increasingly used in many aspects of people's
lives. In real-world applications of planning, multiple optimization objectives
are often involved. Thus, planning agents' decisions can involve complex
tradeoffs among competing objectives. It can be difficult for the end-users to
understand why an agent decides on a particular planning solution on the basis
of its objective values. As a result, the users may not know whether the agent
is making the right decisions, and may lack trust in it. In this work, we
contribute an approach, based on contrastive explanation, that enables a
multi-objective MDP planning agent to explain its decisions in a way that
communicates its tradeoff rationale in terms of the domain-level concepts. We
conduct a human subjects experiment to evaluate the effectiveness of our
explanation approach in a mobile robot navigation domain. The results show that
our approach significantly improves the users' understanding, and confidence in
their understanding, of the tradeoff rationale of the planning agent.
Related papers
- Towards Objective and Unbiased Decision Assessments with LLM-Enhanced Hierarchical Attention Networks [6.520709313101523]
This work investigates cognitive bias identification in high-stake decision making process by human experts.
We propose bias-aware AI-augmented workflow that surpass human judgment.
In our experiments, both the proposed model and the agentic workflow significantly improves on both human judgment and alternative models.
arXiv Detail & Related papers (2024-11-13T10:42:11Z) - Interactive Speculative Planning: Enhance Agent Efficiency through Co-design of System and User Interface [38.76937539085164]
This paper presents a human-centered efficient agent planning method -- Interactive Speculative Planning.
We aim at enhancing the efficiency of agent planning through both system design and human-AI interaction.
arXiv Detail & Related papers (2024-09-30T16:52:51Z) - Ask-before-Plan: Proactive Language Agents for Real-World Planning [68.08024918064503]
Proactive Agent Planning requires language agents to predict clarification needs based on user-agent conversation and agent-environment interaction.
We propose a novel multi-agent framework, Clarification-Execution-Planning (textttCEP), which consists of three agents specialized in clarification, execution, and planning.
arXiv Detail & Related papers (2024-06-18T14:07:28Z) - ADESSE: Advice Explanations in Complex Repeated Decision-Making Environments [14.105935964906976]
This work considers a problem setup where an intelligent agent provides advice to a human decision-maker.
We develop an approach named ADESSE to generate explanations about the adviser agent to improve human trust and decision-making.
arXiv Detail & Related papers (2024-05-31T08:59:20Z) - Tell Me More! Towards Implicit User Intention Understanding of Language
Model Driven Agents [110.25679611755962]
Current language model-driven agents often lack mechanisms for effective user participation, which is crucial given the vagueness commonly found in user instructions.
We introduce Intention-in-Interaction (IN3), a novel benchmark designed to inspect users' implicit intentions through explicit queries.
We empirically train Mistral-Interact, a powerful model that proactively assesses task vagueness, inquires user intentions, and refines them into actionable goals.
arXiv Detail & Related papers (2024-02-14T14:36:30Z) - Rational Decision-Making Agent with Internalized Utility Judgment [91.80700126895927]
Large language models (LLMs) have demonstrated remarkable advancements and have attracted significant efforts to develop LLMs into agents capable of executing intricate multi-step decision-making tasks beyond traditional NLP applications.
This paper proposes RadAgent, which fosters the development of its rationality through an iterative framework involving Experience Exploration and Utility Learning.
Experimental results on the ToolBench dataset demonstrate RadAgent's superiority over baselines, achieving over 10% improvement in Pass Rate on diverse tasks.
arXiv Detail & Related papers (2023-08-24T03:11:45Z) - Formalizing the Problem of Side Effect Regularization [81.97441214404247]
We propose a formal criterion for side effect regularization via the assistance game framework.
In these games, the agent solves a partially observable Markov decision process.
We show that this POMDP is solved by trading off the proxy reward with the agent's ability to achieve a range of future tasks.
arXiv Detail & Related papers (2022-06-23T16:36:13Z) - Reinforcement Learning with a Terminator [80.34572413850186]
We learn the parameters of the TerMDP and leverage the structure of the estimation problem to provide state-wise confidence bounds.
We use these to construct a provably-efficient algorithm, which accounts for termination, and bound its regret.
arXiv Detail & Related papers (2022-05-30T18:40:28Z) - Inverse Online Learning: Understanding Non-Stationary and Reactionary
Policies [79.60322329952453]
We show how to develop interpretable representations of how agents make decisions.
By understanding the decision-making processes underlying a set of observed trajectories, we cast the policy inference problem as the inverse to this online learning problem.
We introduce a practical algorithm for retrospectively estimating such perceived effects, alongside the process through which agents update them.
Through application to the analysis of UNOS organ donation acceptance decisions, we demonstrate that our approach can bring valuable insights into the factors that govern decision processes and how they change over time.
arXiv Detail & Related papers (2022-03-14T17:40:42Z) - Efficient Multi-agent Epistemic Planning: Teaching Planners About Nested
Belief [27.524600740450126]
We plan from the perspective of a single agent with the potential for goals and actions that involve nested beliefs, non-homogeneous agents, co-present observations, and the ability for one agent to reason as if it were another.
Our approach represents an important step towards applying the well-established field of automated planning to the challenging task of planning involving nested beliefs of multiple agents.
arXiv Detail & Related papers (2021-10-06T03:24:01Z) - Comprehensive Multi-Agent Epistemic Planning [0.0]
This manuscript is focused on a specialized kind of planning known as Multi-agent Epistemic Planning (MEP).
EP refers to an automated planning setting where the agent reasons in the space of knowledge/beliefs states and tries to find a plan to reach a desirable state from a starting one.
Its general form, the MEP problem, involves multiple agents who need to reason about both the state of the world and the information flows between agents.
arXiv Detail & Related papers (2021-09-17T01:50:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.