A model-based framework for learning transparent swarm behaviors
- URL: http://arxiv.org/abs/2103.05343v1
- Date: Tue, 9 Mar 2021 10:45:57 GMT
- Title: A model-based framework for learning transparent swarm behaviors
- Authors: Mario Coppola, Jian Guo, Eberhard Gill, Guido C. H. E. de Croon
- Abstract summary: This paper proposes a model-based framework to design understandable and verifiable behaviors for swarms of robots.
The framework is tested on four case studies, featuring aggregation and foraging tasks.
- Score: 6.310689648471231
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes a model-based framework to automatically and efficiently
design understandable and verifiable behaviors for swarms of robots. The
framework is based on the automatic extraction of two distinct models: 1) a
neural network model trained to estimate the relationship between the robots'
sensor readings and the global performance of the swarm, and 2) a probabilistic
state transition model that explicitly models the local state transitions
(i.e., transitions in observations from the perspective of a single robot in
the swarm) given a policy. The models can be trained from a data set of
simulated runs featuring random policies. The first model is used to
automatically extract a set of local states that are expected to maximize the
global performance. These local states are referred to as desired local states.
The second model is used to optimize a stochastic policy so as to increase the
probability that the robots in the swarm observe one of the desired local
states. Following these steps, the framework proposed in this paper can
efficiently lead to effective controllers. This is tested on four case studies,
featuring aggregation and foraging tasks. Importantly, thanks to the models,
the framework allows us to understand and inspect a swarm's behavior. To this
end, we propose verification checks to identify some potential issues that may
prevent the swarm from achieving the desired global objective. In addition, we
explore how the framework can be used in combination with a "standard"
evolutionary robotics strategy (i.e., where performance is measured via
simulation), or with online learning.
Related papers
- Fine-Tuning Generative Models as an Inference Method for Robotic Tasks [18.745665662647912]
We investigate how to quickly adapt the sample generation of neural network models to observations in robotic tasks.
The key idea is to quickly fine-tune the model by fitting it to generated samples matching the observed evidence.
We show that our method can be applied to both autoregressive models and variational autoencoders.
arXiv Detail & Related papers (2023-10-19T16:11:49Z) - NoMaD: Goal Masked Diffusion Policies for Navigation and Exploration [57.15811390835294]
This paper describes how we can train a single unified diffusion policy to handle both goal-directed navigation and goal-agnostic exploration.
We show that this unified policy results in better overall performance when navigating to visually indicated goals in novel environments.
Our experiments, conducted on a real-world mobile robot platform, show effective navigation in unseen environments in comparison with five alternative methods.
arXiv Detail & Related papers (2023-10-11T21:07:14Z) - From system models to class models: An in-context learning paradigm [0.0]
We introduce a novel paradigm for system identification, addressing two primary tasks: one-step-ahead prediction and multi-step simulation.
We learn a meta model that represents a class of dynamical systems.
For one-step prediction, a GPT-like decoder-only architecture is utilized, whereas the simulation problem employs an encoder-decoder structure.
arXiv Detail & Related papers (2023-08-25T13:50:17Z) - RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic
Control [140.48218261864153]
We study how vision-language models trained on Internet-scale data can be incorporated directly into end-to-end robotic control.
Our approach leads to performant robotic policies and enables RT-2 to obtain a range of emergent capabilities from Internet-scale training.
arXiv Detail & Related papers (2023-07-28T21:18:02Z) - Locally adaptive cellular automata for goal-oriented self-organization [14.059479351946386]
We propose a new model class of adaptive cellular automata that allows for the generation of scalable and expressive models.
We show how to implement adaptation by coupling the update rule of the cellular automaton with itself and the system state in a localized way.
arXiv Detail & Related papers (2023-06-12T12:32:23Z) - STMT: A Spatial-Temporal Mesh Transformer for MoCap-Based Action Recognition [50.064502884594376]
We study the problem of human action recognition using motion capture (MoCap) sequences.
We propose a novel Spatial-Temporal Mesh Transformer (STMT) to directly model the mesh sequences.
The proposed method achieves state-of-the-art performance compared to skeleton-based and point-cloud-based models.
arXiv Detail & Related papers (2023-03-31T16:19:27Z) - TrafficBots: Towards World Models for Autonomous Driving Simulation and
Motion Prediction [149.5716746789134]
We show data-driven traffic simulation can be formulated as a world model.
We present TrafficBots, a multi-agent policy built upon motion prediction and end-to-end driving.
Experiments on the open motion dataset show TrafficBots can simulate realistic multi-agent behaviors.
arXiv Detail & Related papers (2023-03-07T18:28:41Z) - IR-MCL: Implicit Representation-Based Online Global Localization [31.77645160411745]
In this paper, we address the problem of estimating the robots pose in an indoor environment using 2D LiDAR data.
We propose a neural occupancy field (NOF) to implicitly represent the scene using a neural network.
We show that we can accurately and efficiently localize a robot using our approach surpassing the localization performance of state-of-the-art methods.
arXiv Detail & Related papers (2022-10-06T17:59:08Z) - Value-Consistent Representation Learning for Data-Efficient
Reinforcement Learning [105.70602423944148]
We propose a novel method, called value-consistent representation learning (VCR), to learn representations that are directly related to decision-making.
Instead of aligning this imagined state with a real state returned by the environment, VCR applies a $Q$-value head on both states and obtains two distributions of action values.
It has been demonstrated that our methods achieve new state-of-the-art performance for search-free RL algorithms.
arXiv Detail & Related papers (2022-06-25T03:02:25Z) - SA-Det3D: Self-Attention Based Context-Aware 3D Object Detection [9.924083358178239]
We propose two variants of self-attention for contextual modeling in 3D object detection.
We first incorporate the pairwise self-attention mechanism into the current state-of-the-art BEV, voxel and point-based detectors.
Next, we propose a self-attention variant that samples a subset of the most representative features by learning deformations over randomly sampled locations.
arXiv Detail & Related papers (2021-01-07T18:30:32Z) - Learning Meta Face Recognition in Unseen Domains [74.69681594452125]
We propose a novel face recognition method via meta-learning named Meta Face Recognition (MFR)
MFR synthesizes the source/target domain shift with a meta-optimization objective.
We propose two benchmarks for generalized face recognition evaluation.
arXiv Detail & Related papers (2020-03-17T14:10:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.