AdverSAR: Adversarial Search and Rescue via Multi-Agent Reinforcement
Learning
- URL: http://arxiv.org/abs/2212.10064v1
- Date: Tue, 20 Dec 2022 08:13:29 GMT
- Title: AdverSAR: Adversarial Search and Rescue via Multi-Agent Reinforcement
Learning
- Authors: Aowabin Rahman, Arnab Bhattacharya, Thiagarajan Ramachandran, Sayak
Mukherjee, Himanshu Sharma, Ted Fujimoto, Samrat Chatterjee
- Abstract summary: We propose an algorithm that allows robots to efficiently coordinate their strategies in the presence of adversarial inter-agent communications.
It is assumed that the robots have no prior knowledge of the target locations, and they can interact with only a subset of neighboring robots at any time.
The effectiveness of our approach is demonstrated on a collection of prototype grid-world environments.
- Score: 4.843554492319537
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Search and Rescue (SAR) missions in remote environments often employ
autonomous multi-robot systems that learn, plan, and execute a combination of
local single-robot control actions, group primitives, and global
mission-oriented coordination and collaboration. Often, SAR coordination
strategies are manually designed by human experts who can remotely control the
multi-robot system and enable semi-autonomous operations. However, in remote
environments where connectivity is limited and human intervention is often not
possible, decentralized collaboration strategies are needed for
fully-autonomous operations. Nevertheless, decentralized coordination may be
ineffective in adversarial environments due to sensor noise, actuation faults,
or manipulation of inter-agent communication data. In this paper, we propose an
algorithmic approach based on adversarial multi-agent reinforcement learning
(MARL) that allows robots to efficiently coordinate their strategies in the
presence of adversarial inter-agent communications. In our setup, the objective
of the multi-robot team is to discover targets strategically in an
obstacle-strewn geographical area by minimizing the average time needed to find
the targets. It is assumed that the robots have no prior knowledge of the
target locations, and they can interact with only a subset of neighboring
robots at any time. Based on the centralized training with decentralized
execution (CTDE) paradigm in MARL, we utilize a hierarchical meta-learning
framework to learn dynamic team-coordination modalities and discover emergent
team behavior under complex cooperative-competitive scenarios. The
effectiveness of our approach is demonstrated on a collection of prototype
grid-world environments with different specifications of benign and adversarial
agents, target locations, and agent rewards.
Related papers
- Decentralized and Lifelong-Adaptive Multi-Agent Collaborative Learning [57.652899266553035]
Decentralized and lifelong-adaptive multi-agent collaborative learning aims to enhance collaboration among multiple agents without a central server.
We propose DeLAMA, a decentralized multi-agent lifelong collaborative learning algorithm with dynamic collaboration graphs.
arXiv Detail & Related papers (2024-03-11T09:21:11Z) - LPAC: Learnable Perception-Action-Communication Loops with Applications
to Coverage Control [80.86089324742024]
We propose a learnable Perception-Action-Communication (LPAC) architecture for the problem.
CNN processes localized perception; a graph neural network (GNN) facilitates robot communications.
Evaluations show that the LPAC models outperform standard decentralized and centralized coverage control algorithms.
arXiv Detail & Related papers (2024-01-10T00:08:00Z) - Asynchronous Perception-Action-Communication with Graph Neural Networks [93.58250297774728]
Collaboration in large robot swarms to achieve a common global objective is a challenging problem in large environments.
The robots must execute a Perception-Action-Communication loop -- they perceive their local environment, communicate with other robots, and take actions in real time.
Recently, this has been addressed using Graph Neural Networks (GNNs) for applications such as flocking and coverage control.
This paper proposes a framework for asynchronous PAC in robot swarms, where decentralized GNNs are used to compute navigation actions and generate messages for communication.
arXiv Detail & Related papers (2023-09-18T21:20:50Z) - Communication-Efficient Reinforcement Learning in Swarm Robotic Networks
for Maze Exploration [2.958532752589616]
Communication is key to the successful coordination of swarm robots.
This paper proposes a new communication-efficient decentralized cooperative reinforcement learning algorithm for coordinating swarm robots.
arXiv Detail & Related papers (2023-05-26T16:56:00Z) - Efficient Skill Acquisition for Complex Manipulation Tasks in Obstructed
Environments [18.348489257164356]
We propose a system for efficient skill acquisition that leverages an object-centric generative model (OCGM) for versatile goal identification.
OCGM enables one-shot target object identification and re-identification in new scenes, allowing MP to guide the robot to the target object while avoiding obstacles.
arXiv Detail & Related papers (2023-03-06T18:49:59Z) - Simulation of robot swarms for learning communication-aware coordination [0.0]
We train end-to-end Neural Networks that take as input local observations obtained from an omniscient centralised controller.
Experiments are run in Enki, a high-performance open-source simulator for planar robots.
arXiv Detail & Related papers (2023-02-25T17:17:40Z) - Hierarchical Reinforcement Learning with Opponent Modeling for
Distributed Multi-agent Cooperation [13.670618752160594]
Deep reinforcement learning (DRL) provides a promising approach for multi-agent cooperation through the interaction of the agents and environments.
Traditional DRL solutions suffer from the high dimensions of multiple agents with continuous action space during policy search.
We propose a hierarchical reinforcement learning approach with high-level decision-making and low-level individual control for efficient policy search.
arXiv Detail & Related papers (2022-06-25T19:09:29Z) - SABER: Data-Driven Motion Planner for Autonomously Navigating
Heterogeneous Robots [112.2491765424719]
We present an end-to-end online motion planning framework that uses a data-driven approach to navigate a heterogeneous robot team towards a global goal.
We use model predictive control (SMPC) to calculate control inputs that satisfy robot dynamics, and consider uncertainty during obstacle avoidance with chance constraints.
recurrent neural networks are used to provide a quick estimate of future state uncertainty considered in the SMPC finite-time horizon solution.
A Deep Q-learning agent is employed to serve as a high-level path planner, providing the SMPC with target positions that move the robots towards a desired global goal.
arXiv Detail & Related papers (2021-08-03T02:56:21Z) - Graph Neural Networks for Decentralized Multi-Robot Submodular Action
Selection [101.38634057635373]
We focus on applications where robots are required to jointly select actions to maximize team submodular objectives.
We propose a general-purpose learning architecture towards submodular at scale, with decentralized communications.
We demonstrate the performance of our GNN-based learning approach in a scenario of active target coverage with large networks of robots.
arXiv Detail & Related papers (2021-05-18T15:32:07Z) - Learning Connectivity for Data Distribution in Robot Teams [96.39864514115136]
We propose a task-agnostic, decentralized, low-latency method for data distribution in ad-hoc networks using Graph Neural Networks (GNN)
Our approach enables multi-agent algorithms based on global state information to function by ensuring it is available at each robot.
We train the distributed GNN communication policies via reinforcement learning using the average Age of Information as the reward function and show that it improves training stability compared to task-specific reward functions.
arXiv Detail & Related papers (2021-03-08T21:48:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.