Scalable Task-Driven Robotic Swarm Control via Collision Avoidance and
Learning Mean-Field Control
- URL: http://arxiv.org/abs/2209.07420v1
- Date: Thu, 15 Sep 2022 16:15:04 GMT
- Title: Scalable Task-Driven Robotic Swarm Control via Collision Avoidance and
Learning Mean-Field Control
- Authors: Kai Cui, Mengguang Li, Christian Fabian, Heinz Koeppl
- Abstract summary: We use state-of-the-art mean-field control techniques to convert many-agent swarm control into classical single-agent control of distributions.
Here, we combine collision avoidance and learning of mean-field control into a unified framework for tractably designing intelligent robotic swarm behavior.
- Score: 23.494528616672024
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, reinforcement learning and its multi-agent analogue have
achieved great success in solving various complex control problems. However,
multi-agent reinforcement learning remains challenging both in its theoretical
analysis and empirical design of algorithms, especially for large swarms of
embodied robotic agents where a definitive toolchain remains part of active
research. We use emerging state-of-the-art mean-field control techniques in
order to convert many-agent swarm control into more classical single-agent
control of distributions. This allows profiting from advances in single-agent
reinforcement learning at the cost of assuming weak interaction between agents.
As a result, the mean-field model is violated by the nature of real systems
with embodied, physically colliding agents. Here, we combine collision
avoidance and learning of mean-field control into a unified framework for
tractably designing intelligent robotic swarm behavior. On the theoretical
side, we provide novel approximation guarantees for both general mean-field
control in continuous spaces and with collision avoidance. On the practical
side, we show that our approach outperforms multi-agent reinforcement learning
and allows for decentralized open-loop application while avoiding collisions,
both in simulation and real UAV swarms. Overall, we propose a framework for the
design of swarm behavior that is both mathematically well-founded and
practically useful, enabling the solution of otherwise intractable swarm
problems.
Related papers
- Decentralized Adversarial Training over Graphs [55.28669771020857]
The vulnerability of machine learning models to adversarial attacks has been attracting considerable attention in recent years.
This work studies adversarial training over graphs, where individual agents are subjected to varied strength perturbation space.
arXiv Detail & Related papers (2023-03-23T15:05:16Z) - Safe Multi-agent Learning via Trapping Regions [89.24858306636816]
We apply the concept of trapping regions, known from qualitative theory of dynamical systems, to create safety sets in the joint strategy space for decentralized learning.
We propose a binary partitioning algorithm for verification that candidate sets form trapping regions in systems with known learning dynamics, and a sampling algorithm for scenarios where learning dynamics are not known.
arXiv Detail & Related papers (2023-02-27T14:47:52Z) - Coach-assisted Multi-Agent Reinforcement Learning Framework for
Unexpected Crashed Agents [120.91291581594773]
We present a formal formulation of a cooperative multi-agent reinforcement learning system with unexpected crashes.
We propose a coach-assisted multi-agent reinforcement learning framework, which introduces a virtual coach agent to adjust the crash rate during training.
To the best of our knowledge, this work is the first to study the unexpected crashes in the multi-agent system.
arXiv Detail & Related papers (2022-03-16T08:22:45Z) - Relative Distributed Formation and Obstacle Avoidance with Multi-agent
Reinforcement Learning [20.401609420707867]
We propose a distributed formation and obstacle avoidance method based on multi-agent reinforcement learning (MARL)
Our method achieves better performance regarding formation error, formation convergence rate and on-par success rate of obstacle avoidance compared with baselines.
arXiv Detail & Related papers (2021-11-14T13:02:45Z) - Learning to Centralize Dual-Arm Assembly [0.6091702876917281]
This work focuses on assembly with humanoid robots by providing a framework for dual-arm peg-in-hole manipulation.
We reduce modeling effort to a minimum by using sparse rewards only.
We demonstrate the effectiveness of the framework on dual-arm peg-in-hole and analyze sample efficiency and success rates for different action spaces.
arXiv Detail & Related papers (2021-10-08T09:59:12Z) - Efficient Model-Based Multi-Agent Mean-Field Reinforcement Learning [89.31889875864599]
We propose an efficient model-based reinforcement learning algorithm for learning in multi-agent systems.
Our main theoretical contributions are the first general regret bounds for model-based reinforcement learning for MFC.
We provide a practical parametrization of the core optimization problem.
arXiv Detail & Related papers (2021-07-08T18:01:02Z) - Decentralized Swarm Collision Avoidance for Quadrotors via End-to-End
Reinforcement Learning [28.592704336574158]
We draw biological inspiration from flocks of starlings and apply the insight to end-to-end learned decentralized collision avoidance.
We propose a new, scalable observation model following a biomimetic topological interaction rule.
Our learned policies are tested in simulation and subsequently transferred to real-world drones to validate their real-world applicability.
arXiv Detail & Related papers (2021-04-30T11:19:03Z) - Discrete-Time Mean Field Control with Environment States [25.44061731738579]
Mean field control and mean field games have been established as a tractable solution for large-scale multi-agent problems with many agents.
We rigorously establish approximate optimality as the number of agents grows in the finite agent case.
We find that a dynamic programming principle holds, resulting in the existence of an optimal stationary policy.
arXiv Detail & Related papers (2021-04-30T10:58:01Z) - UPDeT: Universal Multi-agent Reinforcement Learning via Policy
Decoupling with Transformers [108.92194081987967]
We make the first attempt to explore a universal multi-agent reinforcement learning pipeline, designing one single architecture to fit tasks.
Unlike previous RNN-based models, we utilize a transformer-based model to generate a flexible policy.
The proposed model, named as Universal Policy Decoupling Transformer (UPDeT), further relaxes the action restriction and makes the multi-agent task's decision process more explainable.
arXiv Detail & Related papers (2021-01-20T07:24:24Z) - Learning Safe Multi-Agent Control with Decentralized Neural Barrier
Certificates [19.261536710315028]
We study the multi-agent safe control problem where agents should avoid collisions to static obstacles and collisions with each other while reaching their goals.
Our core idea is to learn the multi-agent control policy jointly with learning the control barrier functions as safety certificates.
We propose a novel joint-learning framework that can be implemented in a decentralized fashion, with generalization guarantees for certain function classes.
arXiv Detail & Related papers (2021-01-14T03:17:17Z) - Risk-Sensitive Sequential Action Control with Multi-Modal Human
Trajectory Forecasting for Safe Crowd-Robot Interaction [55.569050872780224]
We present an online framework for safe crowd-robot interaction based on risk-sensitive optimal control, wherein the risk is modeled by the entropic risk measure.
Our modular approach decouples the crowd-robot interaction into learning-based prediction and model-based control.
A simulation study and a real-world experiment show that the proposed framework can accomplish safe and efficient navigation while avoiding collisions with more than 50 humans in the scene.
arXiv Detail & Related papers (2020-09-12T02:02:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.