MUI-TARE: Multi-Agent Cooperative Exploration with Unknown Initial
Position
- URL: http://arxiv.org/abs/2209.10775v1
- Date: Thu, 22 Sep 2022 04:33:02 GMT
- Title: MUI-TARE: Multi-Agent Cooperative Exploration with Unknown Initial
Position
- Authors: Jingtian Yan, Xingqiao Lin, Zhongqiang Ren, Shiqi Zhao, Jieqiong Yu,
Chao Cao, Peng Yin, Ji Zhang, and Sebastian Scherer
- Abstract summary: We develop a new approach for lidar-based multi-agent exploration based on the quality indicator of the sub-map merging process.
Our approach is up to 50% more efficient than the baselines on average while merging sub-maps robustly.
- Score: 12.921108151387696
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-agent exploration of a bounded 3D environment with unknown initial
positions of agents is a challenging problem. It requires quickly exploring the
environments as well as robustly merging the sub-maps built by the agents. We
take the view that the existing approaches are either aggressive or
conservative: Aggressive strategies merge two sub-maps built by different
agents together when overlap is detected, which can lead to incorrect merging
due to the false-positive detection of the overlap and is thus not robust.
Conservative strategies direct one agent to revisit an excessive amount of the
historical trajectory of another agent for verification before merging, which
can lower the exploration efficiency due to the repeated exploration of the
same space. To intelligently balance the robustness of sub-map merging and
exploration efficiency, we develop a new approach for lidar-based multi-agent
exploration, which can direct one agent to repeat another agent's trajectory in
an \emph{adaptive} manner based on the quality indicator of the sub-map merging
process. Additionally, our approach extends the recent single-agent
hierarchical exploration strategy to multiple agents in a \emph{cooperative}
manner by planning for agents with merged sub-maps together to further improve
exploration efficiency. Our experiments show that our approach is up to 50\%
more efficient than the baselines on average while merging sub-maps robustly.
Related papers
- Decentralized Monte Carlo Tree Search for Partially Observable
Multi-agent Pathfinding [49.730902939565986]
Multi-Agent Pathfinding problem involves finding a set of conflict-free paths for a group of agents confined to a graph.
In this study, we focus on the decentralized MAPF setting, where the agents may observe the other agents only locally.
We propose a decentralized multi-agent Monte Carlo Tree Search (MCTS) method for MAPF tasks.
arXiv Detail & Related papers (2023-12-26T06:57:22Z) - Deep Multi-Agent Reinforcement Learning for Decentralized Active
Hypothesis Testing [11.639503711252663]
We tackle the multi-agent active hypothesis testing (AHT) problem by introducing a novel algorithm rooted in the framework of deep multi-agent reinforcement learning.
We present a comprehensive set of experimental results that effectively showcase the agents' ability to learn collaborative strategies and enhance performance.
arXiv Detail & Related papers (2023-09-14T01:18:04Z) - Scalable Multi-agent Covering Option Discovery based on Kronecker Graphs [49.71319907864573]
In this paper, we propose multi-agent skill discovery which enables the ease of decomposition.
Our key idea is to approximate the joint state space as a Kronecker graph, based on which we can directly estimate its Fiedler vector.
Considering that directly computing the Laplacian spectrum is intractable for tasks with infinite-scale state spaces, we further propose a deep learning extension of our method.
arXiv Detail & Related papers (2023-07-21T14:53:12Z) - Multi-agent Deep Covering Skill Discovery [50.812414209206054]
We propose Multi-agent Deep Covering Option Discovery, which constructs the multi-agent options through minimizing the expected cover time of the multiple agents' joint state space.
Also, we propose a novel framework to adopt the multi-agent options in the MARL process.
We show that the proposed algorithm can effectively capture the agent interactions with the attention mechanism, successfully identify multi-agent options, and significantly outperforms prior works using single-agent options or no options.
arXiv Detail & Related papers (2022-10-07T00:40:59Z) - Learning Multi-agent Skills for Tabular Reinforcement Learning using
Factor Graphs [41.17714498464354]
We show that it is possible to directly compute multi-agent options with collaborative exploratory behaviors among the agents.
The proposed algorithm can successfully identify multi-agent options, and significantly outperforms prior works using single-agent options or no options.
arXiv Detail & Related papers (2022-01-20T15:33:08Z) - Cooperative Exploration for Multi-Agent Deep Reinforcement Learning [127.4746863307944]
We propose cooperative multi-agent exploration (CMAE) for deep reinforcement learning.
The goal is selected from multiple projected state spaces via a normalized entropy-based technique.
We demonstrate that CMAE consistently outperforms baselines on various tasks.
arXiv Detail & Related papers (2021-07-23T20:06:32Z) - Decentralised Approach for Multi Agent Path Finding [6.599344783327053]
Multi Agent Path Finding (MAPF) requires identification of conflict free paths for spatially-extended agents.
These find application in real world problems like Convoy Movement Problem, Train Scheduling etc.
Our proposed approach, Decentralised Multi Agent Path Finding (DeMAPF), handles MAPF as a sequence of pathplanning and allocation problems.
arXiv Detail & Related papers (2021-06-03T18:07:26Z) - Scalable Multi-Agent Inverse Reinforcement Learning via
Actor-Attention-Critic [54.2180984002807]
Multi-agent adversarial inverse reinforcement learning (MA-AIRL) is a recent approach that applies single-agent AIRL to multi-agent problems.
We propose a multi-agent inverse RL algorithm that is more sample-efficient and scalable than previous works.
arXiv Detail & Related papers (2020-02-24T20:30:45Z) - Never Give Up: Learning Directed Exploration Strategies [63.19616370038824]
We propose a reinforcement learning agent to solve hard exploration games by learning a range of directed exploratory policies.
We construct an episodic memory-based intrinsic reward using k-nearest neighbors over the agent's recent experience to train the directed exploratory policies.
A self-supervised inverse dynamics model is used to train the embeddings of the nearest neighbour lookup, biasing the novelty signal towards what the agent can control.
arXiv Detail & Related papers (2020-02-14T13:57:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.