Enhanced method for reinforcement learning based dynamic obstacle
avoidance by assessment of collision risk
- URL: http://arxiv.org/abs/2212.04123v1
- Date: Thu, 8 Dec 2022 07:46:42 GMT
- Title: Enhanced method for reinforcement learning based dynamic obstacle
avoidance by assessment of collision risk
- Authors: Fabian Hart, Ostap Okhrin
- Abstract summary: This paper proposes a general training environment where we gain control over the difficulty of the obstacle avoidance task.
We found that shifting the training towards a greater task difficulty can massively increase the final performance.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the field of autonomous robots, reinforcement learning (RL) is an
increasingly used method to solve the task of dynamic obstacle avoidance for
mobile robots, autonomous ships, and drones. A common practice to train those
agents is to use a training environment with random initialization of agent and
obstacles. Such approaches might suffer from a low coverage of high-risk
scenarios in training, leading to impaired final performance of obstacle
avoidance. This paper proposes a general training environment where we gain
control over the difficulty of the obstacle avoidance task by using short
training episodes and assessing the difficulty by two metrics: The number of
obstacles and a collision risk metric. We found that shifting the training
towards a greater task difficulty can massively increase the final performance.
A baseline agent, using a traditional training environment based on random
initialization of agent and obstacles and longer training episodes, leads to a
significantly weaker performance. To prove the generalizability of the proposed
approach, we designed two realistic use cases: A mobile robot and a maritime
ship under the threat of approaching obstacles. In both applications, the
previous results can be confirmed, which emphasizes the general usability of
the proposed approach, detached from a specific application context and
independent of the agent's dynamics. We further added Gaussian noise to the
sensor signals, resulting in only a marginal degradation of performance and
thus indicating solid robustness of the trained agent.
Related papers
- RACER: Epistemic Risk-Sensitive RL Enables Fast Driving with Fewer Crashes [57.319845580050924]
We propose a reinforcement learning framework that combines risk-sensitive control with an adaptive action space curriculum.
We show that our algorithm is capable of learning high-speed policies for a real-world off-road driving task.
arXiv Detail & Related papers (2024-05-07T23:32:36Z) - Multi-granular Adversarial Attacks against Black-box Neural Ranking Models [111.58315434849047]
We create high-quality adversarial examples by incorporating multi-granular perturbations.
We transform the multi-granular attack into a sequential decision-making process.
Our attack method surpasses prevailing baselines in both attack effectiveness and imperceptibility.
arXiv Detail & Related papers (2024-04-02T02:08:29Z) - Towards Deviation-Robust Agent Navigation via Perturbation-Aware
Contrastive Learning [125.61772424068903]
Vision-and-language navigation (VLN) asks an agent to follow a given language instruction to navigate through a real 3D environment.
We present a model-agnostic training paradigm, called Progressive Perturbation-aware Contrastive Learning (PROPER) to enhance the generalization ability of existing VLN agents.
arXiv Detail & Related papers (2024-03-09T02:34:13Z) - Two-step dynamic obstacle avoidance [0.0]
This paper proposes a two-step architecture for handling dynamic obstacle avoidance (DOA) tasks by combining supervised and reinforcement learning (RL)
In the first step, we introduce a data-driven approach to estimate the collision risk (CR) of an obstacle using a recurrent neural network.
In the second step, we include these CR estimates into the observation space of an RL agent to increase its situational awareness.
arXiv Detail & Related papers (2023-11-28T14:55:50Z) - Collision Avoidance and Navigation for a Quadrotor Swarm Using End-to-end Deep Reinforcement Learning [8.864432196281268]
We propose an end-to-end DRL approach to control quadrotor swarms in environments with obstacles.
We provide our agents a curriculum and a replay buffer of the clipped collision episodes to improve performance in obstacle-rich environments.
Our work is the first work that demonstrates the possibility of learning neighbor-avoiding and obstacle-avoiding control policies trained with end-to-end DRL.
arXiv Detail & Related papers (2023-09-23T06:56:28Z) - Adversarial Training Should Be Cast as a Non-Zero-Sum Game [121.95628660889628]
Two-player zero-sum paradigm of adversarial training has not engendered sufficient levels of robustness.
We show that the commonly used surrogate-based relaxation used in adversarial training algorithms voids all guarantees on robustness.
A novel non-zero-sum bilevel formulation of adversarial training yields a framework that matches and in some cases outperforms state-of-the-art attacks.
arXiv Detail & Related papers (2023-06-19T16:00:48Z) - Semi-supervised Semantics-guided Adversarial Training for Trajectory
Prediction [15.707419899141698]
Adversarial attacks on trajectory prediction may mislead the prediction of future trajectories and induce unsafe planning.
We present a novel adversarial training method for trajectory prediction.
Our method can effectively mitigate the impact of adversarial attacks by up to 73% and outperform other popular defense methods.
arXiv Detail & Related papers (2022-05-27T20:50:36Z) - Robust Reinforcement Learning via Genetic Curriculum [5.421464476555662]
Genetic curriculum is an algorithm that automatically identifies scenarios in which the agent currently fails and generates an associated curriculum.
Our empirical studies show improvement in robustness over the existing state of the art algorithms, providing training curricula that result in agents being 2 - 8x times less likely to fail.
arXiv Detail & Related papers (2022-02-17T01:14:20Z) - Relative Distributed Formation and Obstacle Avoidance with Multi-agent
Reinforcement Learning [20.401609420707867]
We propose a distributed formation and obstacle avoidance method based on multi-agent reinforcement learning (MARL)
Our method achieves better performance regarding formation error, formation convergence rate and on-par success rate of obstacle avoidance compared with baselines.
arXiv Detail & Related papers (2021-11-14T13:02:45Z) - Adversarial Training is Not Ready for Robot Learning [55.493354071227174]
Adversarial training is an effective method to train deep learning models that are resilient to norm-bounded perturbations.
We show theoretically and experimentally that neural controllers obtained via adversarial training are subjected to three types of defects.
Our results suggest that adversarial training is not yet ready for robot learning.
arXiv Detail & Related papers (2021-03-15T07:51:31Z) - Robust Reinforcement Learning using Adversarial Populations [118.73193330231163]
Reinforcement Learning (RL) is an effective tool for controller design but can struggle with issues of robustness.
We show that using a single adversary does not consistently yield robustness to dynamics variations under standard parametrizations of the adversary.
We propose a population-based augmentation to the Robust RL formulation in which we randomly initialize a population of adversaries and sample from the population uniformly during training.
arXiv Detail & Related papers (2020-08-04T20:57:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.