Towards Using Promises for Multi-Agent Cooperation in Goal Reasoning
- URL: http://arxiv.org/abs/2206.09864v1
- Date: Mon, 20 Jun 2022 15:57:51 GMT
- Title: Towards Using Promises for Multi-Agent Cooperation in Goal Reasoning
- Authors: Daniel Swoboda, Till Hofmann, Tarik Viehmann, Gerhard Lakemeyer
- Abstract summary: We show how promises can be incorporated into the goal life cycle, a commonly used goal refinement mechanism.
We then show how promises can be used when planning for a particular goal by connecting them to timed initial literals.
- Score: 15.924281804465254
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Reasoning and planning for mobile robots is a challenging problem, as the
world evolves over time and thus the robot's goals may change. One technique to
tackle this problem is goal reasoning, where the agent not only reasons about
its actions, but also about which goals to pursue. While goal reasoning for
single agents has been researched extensively, distributed, multi-agent goal
reasoning comes with additional challenges, especially in a distributed
setting. In such a context, some form of coordination is necessary to allow for
cooperative behavior. Previous goal reasoning approaches share the agent's
world model with the other agents, which already enables basic cooperation.
However, the agent's goals, and thus its intentions, are typically not shared.
In this paper, we present a method to tackle this limitation. Extending an
existing goal reasoning framework, we propose enabling cooperative behavior
between multiple agents through promises, where an agent may promise that
certain facts will be true at some point in the future. Sharing these promises
allows other agents to not only consider the current state of the world, but
also the intentions of other agents when deciding on which goal to pursue next.
We describe how promises can be incorporated into the goal life cycle, a
commonly used goal refinement mechanism. We then show how promises can be used
when planning for a particular goal by connecting them to timed initial
literals (TILs) from PDDL planning. Finally, we evaluate our prototypical
implementation in a simplified logistics scenario.
Related papers
- The Rise and Potential of Large Language Model Based Agents: A Survey [91.71061158000953]
Large language models (LLMs) are regarded as potential sparks for Artificial General Intelligence (AGI)
We start by tracing the concept of agents from its philosophical origins to its development in AI, and explain why LLMs are suitable foundations for agents.
We explore the extensive applications of LLM-based agents in three aspects: single-agent scenarios, multi-agent scenarios, and human-agent cooperation.
arXiv Detail & Related papers (2023-09-14T17:12:03Z) - Adaptation and Communication in Human-Robot Teaming to Handle
Discrepancies in Agents' Beliefs about Plans [13.637799815698559]
We provide an online execution algorithm based on Monte Carlo Tree Search for the agent to plan its action.
We show that our agent is better equipped to work in teams without the guarantee of a shared mental model.
arXiv Detail & Related papers (2023-07-07T03:05:34Z) - Inferring the Goals of Communicating Agents from Actions and
Instructions [47.5816320484482]
We introduce a model of a cooperative team where one agent, the principal, may communicate natural language instructions about their shared plan to another agent, the assistant.
We show how a third person observer can infer the team's goal via multi-modal inverse planning from actions and instructions.
We evaluate this approach by comparing it with human goal inferences in a multi-agent gridworld, finding that our model's inferences closely correlate with human judgments.
arXiv Detail & Related papers (2023-06-28T13:43:46Z) - A Fully Controllable Agent in the Path Planning using Goal-Conditioned
Reinforcement Learning [0.0]
In the path planning, the routes may vary depending on the number of variables such as that it is important for the agent to reach various goals.
I propose a novel reinforcement learning framework for a fully controllable agent in the path planning.
arXiv Detail & Related papers (2022-05-20T05:18:03Z) - Planning to Practice: Efficient Online Fine-Tuning by Composing Goals in
Latent Space [76.46113138484947]
General-purpose robots require diverse repertoires of behaviors to complete challenging tasks in real-world unstructured environments.
To address this issue, goal-conditioned reinforcement learning aims to acquire policies that can reach goals for a wide range of tasks on command.
We propose Planning to Practice, a method that makes it practical to train goal-conditioned policies for long-horizon tasks.
arXiv Detail & Related papers (2022-05-17T06:58:17Z) - Bisimulation Makes Analogies in Goal-Conditioned Reinforcement Learning [71.52722621691365]
Building generalizable goal-conditioned agents from rich observations is a key to reinforcement learning (RL) solving real world problems.
We propose a new form of state abstraction called goal-conditioned bisimulation.
We learn this representation using a metric form of this abstraction, and show its ability to generalize to new goals in simulation manipulation tasks.
arXiv Detail & Related papers (2022-04-27T17:00:11Z) - Efficient Multi-agent Epistemic Planning: Teaching Planners About Nested
Belief [27.524600740450126]
We plan from the perspective of a single agent with the potential for goals and actions that involve nested beliefs, non-homogeneous agents, co-present observations, and the ability for one agent to reason as if it were another.
Our approach represents an important step towards applying the well-established field of automated planning to the challenging task of planning involving nested beliefs of multiple agents.
arXiv Detail & Related papers (2021-10-06T03:24:01Z) - Modelling Multi-Agent Epistemic Planning in ASP [66.76082318001976]
This paper presents an implementation of a multi-shot Answer Set Programming-based planner that can reason in multi-agent epistemic settings.
The paper shows how the planner, exploiting an ad-hoc epistemic state representation and the efficiency of ASP solvers, has competitive performance results on benchmarks collected from the literature.
arXiv Detail & Related papers (2020-08-07T06:35:56Z) - Automatic Curriculum Learning through Value Disagreement [95.19299356298876]
Continually solving new, unsolved tasks is the key to learning diverse behaviors.
In the multi-task domain, where an agent needs to reach multiple goals, the choice of training goals can largely affect sample efficiency.
We propose setting up an automatic curriculum for goals that the agent needs to solve.
We evaluate our method across 13 multi-goal robotic tasks and 5 navigation tasks, and demonstrate performance gains over current state-of-the-art methods.
arXiv Detail & Related papers (2020-06-17T03:58:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.