Scalable Multi-agent Covering Option Discovery based on Kronecker Graphs
- URL: http://arxiv.org/abs/2307.11629v2
- Date: Sun, 20 Aug 2023 14:26:52 GMT
- Title: Scalable Multi-agent Covering Option Discovery based on Kronecker Graphs
- Authors: Jiayu Chen, Jingdi Chen, Tian Lan, Vaneet Aggarwal
- Abstract summary: In this paper, we propose multi-agent skill discovery which enables the ease of decomposition.
Our key idea is to approximate the joint state space as a Kronecker graph, based on which we can directly estimate its Fiedler vector.
Considering that directly computing the Laplacian spectrum is intractable for tasks with infinite-scale state spaces, we further propose a deep learning extension of our method.
- Score: 49.71319907864573
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Covering skill (a.k.a., option) discovery has been developed to improve the
exploration of RL in single-agent scenarios with sparse reward signals, through
connecting the most distant states in the embedding space provided by the
Fiedler vector of the state transition graph. Given that joint state space
grows exponentially with the number of agents in multi-agent systems, existing
researches still relying on single-agent skill discovery either become
prohibitive or fail to directly discover joint skills that improve the
connectivity of the joint state space. In this paper, we propose multi-agent
skill discovery which enables the ease of decomposition. Our key idea is to
approximate the joint state space as a Kronecker graph, based on which we can
directly estimate its Fiedler vector using the Laplacian spectrum of individual
agents' transition graphs. Further, considering that directly computing the
Laplacian spectrum is intractable for tasks with infinite-scale state spaces,
we further propose a deep learning extension of our method by estimating
eigenfunctions through NN-based representation learning techniques. The
evaluation on multi-agent tasks built with simulators like Mujoco, shows that
the proposed algorithm can successfully identify multi-agent skills, and
significantly outperforms the state-of-the-art. Codes are available at:
https://github.itap.purdue.edu/Clan-labs/Scalable_MAOD_via_KP.
Related papers
- CollaMamba: Efficient Collaborative Perception with Cross-Agent Spatial-Temporal State Space Model [12.461378793357705]
Multi-agent collaborative perception fosters a deeper understanding of the environment.
Recent studies on collaborative perception mostly utilize CNNs or Transformers to learn feature representation and fusion in the spatial dimension.
We propose a resource efficient cross-agent spatial-temporal collaborative state space model (SSM), named CollaMamba.
arXiv Detail & Related papers (2024-09-12T02:50:04Z) - Deciphering Movement: Unified Trajectory Generation Model for Multi-Agent [53.637837706712794]
We propose a Unified Trajectory Generation model, UniTraj, that processes arbitrary trajectories as masked inputs.
Specifically, we introduce a Ghost Spatial Masking (GSM) module embedded within a Transformer encoder for spatial feature extraction.
We benchmark three practical sports game datasets, Basketball-U, Football-U, and Soccer-U, for evaluation.
arXiv Detail & Related papers (2024-05-27T22:15:23Z) - Decentralized Multi-Agent Active Search and Tracking when Targets
Outnumber Agents [8.692007892160913]
We propose a decentralized multi-agent, multi-target, simultaneous active search-and-tracking algorithm called DecSTER.
Our proposed algorithm uses a sequential monte carlo implementation of the probability hypothesis density filter for posterior inference combined with Thompson sampling for decentralized multi-agent decision making.
In simulation, we demonstrate that DecSTER is robust to unreliable inter-agent communication and outperforms information-greedy baselines in terms of the Optimal Sub-Pattern Assignment (OSPA) metric.
arXiv Detail & Related papers (2024-01-06T08:10:58Z) - Decentralized Monte Carlo Tree Search for Partially Observable
Multi-agent Pathfinding [49.730902939565986]
Multi-Agent Pathfinding problem involves finding a set of conflict-free paths for a group of agents confined to a graph.
In this study, we focus on the decentralized MAPF setting, where the agents may observe the other agents only locally.
We propose a decentralized multi-agent Monte Carlo Tree Search (MCTS) method for MAPF tasks.
arXiv Detail & Related papers (2023-12-26T06:57:22Z) - Self-Supervised Neuron Segmentation with Multi-Agent Reinforcement
Learning [53.00683059396803]
Mask image model (MIM) has been widely used due to its simplicity and effectiveness in recovering original information from masked images.
We propose a decision-based MIM that utilizes reinforcement learning (RL) to automatically search for optimal image masking ratio and masking strategy.
Our approach has a significant advantage over alternative self-supervised methods on the task of neuron segmentation.
arXiv Detail & Related papers (2023-10-06T10:40:46Z) - Bridging the Domain Gap for Multi-Agent Perception [19.724227909352976]
We propose the first lightweight framework to bridge domain gaps for multi-agent perception.
Our framework consists of a learnable feature resizer to align features in multiple dimensions and a sparse cross-domain transformer for domain adaption.
Our method can effectively bridge the gap for features from different domains and outperform other baseline methods significantly by at least 8% for point-cloud-based 3D object detection.
arXiv Detail & Related papers (2022-10-16T05:29:21Z) - Multi-agent Deep Covering Skill Discovery [50.812414209206054]
We propose Multi-agent Deep Covering Option Discovery, which constructs the multi-agent options through minimizing the expected cover time of the multiple agents' joint state space.
Also, we propose a novel framework to adopt the multi-agent options in the MARL process.
We show that the proposed algorithm can effectively capture the agent interactions with the attention mechanism, successfully identify multi-agent options, and significantly outperforms prior works using single-agent options or no options.
arXiv Detail & Related papers (2022-10-07T00:40:59Z) - Learning Multi-agent Skills for Tabular Reinforcement Learning using
Factor Graphs [41.17714498464354]
We show that it is possible to directly compute multi-agent options with collaborative exploratory behaviors among the agents.
The proposed algorithm can successfully identify multi-agent options, and significantly outperforms prior works using single-agent options or no options.
arXiv Detail & Related papers (2022-01-20T15:33:08Z) - Soft Hierarchical Graph Recurrent Networks for Many-Agent Partially
Observable Environments [9.067091068256747]
We propose a novel network structure called hierarchical graph recurrent network(HGRN) for multi-agent cooperation under partial observability.
Based on the above technologies, we proposed a value-based MADRL algorithm called Soft-HGRN and its actor-critic variant named SAC-HRGN.
arXiv Detail & Related papers (2021-09-05T09:51:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.