Simulation of robot swarms for learning communication-aware coordination
- URL: http://arxiv.org/abs/2302.13124v1
- Date: Sat, 25 Feb 2023 17:17:40 GMT
- Title: Simulation of robot swarms for learning communication-aware coordination
- Authors: Giorgia Adorni
- Abstract summary: We train end-to-end Neural Networks that take as input local observations obtained from an omniscient centralised controller.
Experiments are run in Enki, a high-performance open-source simulator for planar robots.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Robotics research has been focusing on cooperative multi-agent problems,
where agents must work together and communicate to achieve a shared objective.
To tackle this challenge, we explore imitation learning algorithms. These
methods learn a controller by observing demonstrations of an expert, such as
the behaviour of a centralised omniscient controller, which can perceive the
entire environment, including the state and observations of all agents.
Performing tasks with complete knowledge of the state of a system is
relatively easy, but centralised solutions might not be feasible in real
scenarios since agents do not have direct access to the state but only to their
observations. To overcome this issue, we train end-to-end Neural Networks that
take as input local observations obtained from an omniscient centralised
controller, i.e., the agents' sensor readings and the communications received,
producing as output the action to be performed and the communication to be
transmitted.
This study concentrates on two cooperative tasks using a distributed
controller: distributing the robots evenly in space and colouring them based on
their position relative to others. While an explicit exchange of messages
between the agents is required to solve the second task, in the first one, a
communication protocol is unnecessary, although it may increase performance.
The experiments are run in Enki, a high-performance open-source simulator for
planar robots, which provides collision detection and limited physics support
for robots evolving on a flat surface. Moreover, it can simulate groups of
robots hundreds of times faster than real-time.
The results show how applying a communication strategy improves the
performance of the distributed model, letting it decide which actions to take
almost as precisely and quickly as the expert controller.
Related papers
- Human-Agent Joint Learning for Efficient Robot Manipulation Skill Acquisition [48.65867987106428]
We introduce a novel system for joint learning between human operators and robots.
It enables human operators to share control of a robot end-effector with a learned assistive agent.
It reduces the need for human adaptation while ensuring the collected data is of sufficient quality for downstream tasks.
arXiv Detail & Related papers (2024-06-29T03:37:29Z) - LPAC: Learnable Perception-Action-Communication Loops with Applications
to Coverage Control [80.86089324742024]
We propose a learnable Perception-Action-Communication (LPAC) architecture for the problem.
CNN processes localized perception; a graph neural network (GNN) facilitates robot communications.
Evaluations show that the LPAC models outperform standard decentralized and centralized coverage control algorithms.
arXiv Detail & Related papers (2024-01-10T00:08:00Z) - Asynchronous Perception-Action-Communication with Graph Neural Networks [93.58250297774728]
Collaboration in large robot swarms to achieve a common global objective is a challenging problem in large environments.
The robots must execute a Perception-Action-Communication loop -- they perceive their local environment, communicate with other robots, and take actions in real time.
Recently, this has been addressed using Graph Neural Networks (GNNs) for applications such as flocking and coverage control.
This paper proposes a framework for asynchronous PAC in robot swarms, where decentralized GNNs are used to compute navigation actions and generate messages for communication.
arXiv Detail & Related papers (2023-09-18T21:20:50Z) - Communication-Efficient Reinforcement Learning in Swarm Robotic Networks
for Maze Exploration [2.958532752589616]
Communication is key to the successful coordination of swarm robots.
This paper proposes a new communication-efficient decentralized cooperative reinforcement learning algorithm for coordinating swarm robots.
arXiv Detail & Related papers (2023-05-26T16:56:00Z) - AdverSAR: Adversarial Search and Rescue via Multi-Agent Reinforcement
Learning [4.843554492319537]
We propose an algorithm that allows robots to efficiently coordinate their strategies in the presence of adversarial inter-agent communications.
It is assumed that the robots have no prior knowledge of the target locations, and they can interact with only a subset of neighboring robots at any time.
The effectiveness of our approach is demonstrated on a collection of prototype grid-world environments.
arXiv Detail & Related papers (2022-12-20T08:13:29Z) - A Robot Web for Distributed Many-Device Localisation [18.417301483203996]
We show that a distributed network of robots can collaborate to globally localise via efficient ad-hoc peer to peer communication.
We show in simulations with up to 1000 robots interacting in arbitrary patterns that our solution convergently achieves global accuracy as accurate as a non-linear factor graph solver.
arXiv Detail & Related papers (2022-02-07T16:00:25Z) - Graph Neural Networks for Decentralized Multi-Robot Submodular Action
Selection [101.38634057635373]
We focus on applications where robots are required to jointly select actions to maximize team submodular objectives.
We propose a general-purpose learning architecture towards submodular at scale, with decentralized communications.
We demonstrate the performance of our GNN-based learning approach in a scenario of active target coverage with large networks of robots.
arXiv Detail & Related papers (2021-05-18T15:32:07Z) - Learning Connectivity for Data Distribution in Robot Teams [96.39864514115136]
We propose a task-agnostic, decentralized, low-latency method for data distribution in ad-hoc networks using Graph Neural Networks (GNN)
Our approach enables multi-agent algorithms based on global state information to function by ensuring it is available at each robot.
We train the distributed GNN communication policies via reinforcement learning using the average Age of Information as the reward function and show that it improves training stability compared to task-specific reward functions.
arXiv Detail & Related papers (2021-03-08T21:48:55Z) - Learning to Communicate and Correct Pose Errors [75.03747122616605]
We study the setting proposed in V2VNet, where nearby self-driving vehicles jointly perform object detection and motion forecasting in a cooperative manner.
We propose a novel neural reasoning framework that learns to communicate, to estimate potential errors, and to reach a consensus about those errors.
arXiv Detail & Related papers (2020-11-10T18:19:40Z) - Who2com: Collaborative Perception via Learnable Handshake Communication [34.29310680302486]
We propose the problem of collaborative perception, where robots can combine their local observations with those of neighboring agents in a learnable way to improve accuracy on a perception task.
Inspired by networking communication protocols, we propose a multi-stage handshake communication mechanism where the neural network can learn to compress relevant information needed for each stage.
We show that for the semantic segmentation task, our handshake communication method significantly improves accuracy by approximately 20% over decentralized baselines, and is comparable to centralized ones using a quarter of the bandwidth.
arXiv Detail & Related papers (2020-03-21T04:16:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.