Robust Path Following on Rivers Using Bootstrapped Reinforcement
Learning
- URL: http://arxiv.org/abs/2303.15178v1
- Date: Fri, 24 Mar 2023 07:21:27 GMT
- Title: Robust Path Following on Rivers Using Bootstrapped Reinforcement
Learning
- Authors: Niklas Paulig, Ostap Ohkrin
- Abstract summary: This paper develops a Deep Reinforcement Learning (DRL)-agent for navigation and control of autonomous surface vessels (ASV) on inland waterways.
A state-of-the-art bootstrapped Q-learning algorithm in combination with a versatile training environment generator leads to a robust and accurate rudder controller.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper develops a Deep Reinforcement Learning (DRL)-agent for navigation
and control of autonomous surface vessels (ASV) on inland waterways. Spatial
restrictions due to waterway geometry and the resulting challenges, such as
high flow velocities or shallow banks, require controlled and precise movement
of the ASV. A state-of-the-art bootstrapped Q-learning algorithm in combination
with a versatile training environment generator leads to a robust and accurate
rudder controller. To validate our results, we compare the path-following
capabilities of the proposed approach to a vessel-specific PID controller on
real-world river data from the lower- and middle Rhine, indicating that the DRL
algorithm could effectively prove generalizability even in never-seen scenarios
while simultaneously attaining high navigational accuracy.
Related papers
- Aquatic Navigation: A Challenging Benchmark for Deep Reinforcement Learning [53.3760591018817]
We propose a new benchmarking environment for aquatic navigation using recent advances in the integration between game engines and Deep Reinforcement Learning.
Specifically, we focus on PPO, one of the most widely accepted algorithms, and we propose advanced training techniques.
Our empirical evaluation shows that a well-designed combination of these ingredients can achieve promising results.
arXiv Detail & Related papers (2024-05-30T23:20:23Z) - Variational Autoencoders for exteroceptive perception in reinforcement learning-based collision avoidance [0.0]
Deep Reinforcement Learning (DRL) has emerged as a promising control framework.
Current DRL algorithms require disproportionally large computational resources to find near-optimal policies.
This paper presents a comprehensive exploration of our proposed approach in maritime control systems.
arXiv Detail & Related papers (2024-03-31T09:25:28Z) - Towards Deviation-Robust Agent Navigation via Perturbation-Aware
Contrastive Learning [125.61772424068903]
Vision-and-language navigation (VLN) asks an agent to follow a given language instruction to navigate through a real 3D environment.
We present a model-agnostic training paradigm, called Progressive Perturbation-aware Contrastive Learning (PROPER) to enhance the generalization ability of existing VLN agents.
arXiv Detail & Related papers (2024-03-09T02:34:13Z) - DATT: Deep Adaptive Trajectory Tracking for Quadrotor Control [62.24301794794304]
Deep Adaptive Trajectory Tracking (DATT) is a learning-based approach that can precisely track arbitrary, potentially infeasible trajectories in the presence of large disturbances in the real world.
DATT significantly outperforms competitive adaptive nonlinear and model predictive controllers for both feasible smooth and infeasible trajectories in unsteady wind fields.
It can efficiently run online with an inference time less than 3.2 ms, less than 1/4 of the adaptive nonlinear model predictive control baseline.
arXiv Detail & Related papers (2023-10-13T12:22:31Z) - 2-Level Reinforcement Learning for Ships on Inland Waterways: Path Planning and Following [0.0]
This paper proposes a realistic modularized framework for controlling autonomous surface vehicles (ASVs) on inland waterways (IWs) based on deep reinforcement learning (DRL)
The framework improves operational safety and comprises two levels: a high-level local path planning (LPP) unit and a low-level path following (PF) unit, each consisting of a DRL agent.
arXiv Detail & Related papers (2023-07-25T08:42:59Z) - Efficient Deep Reinforcement Learning Requires Regulating Overfitting [91.88004732618381]
We show that high temporal-difference (TD) error on the validation set of transitions is the main culprit that severely affects the performance of deep RL algorithms.
We show that a simple online model selection method that targets the validation TD error is effective across state-based DMC and Gym tasks.
arXiv Detail & Related papers (2023-04-20T17:11:05Z) - Using Deep Reinforcement Learning with Automatic Curriculum earning for
Mapless Navigation in Intralogistics [0.7633618497843278]
We propose a deep reinforcement learning approach for solving a mapless navigation problem in warehouse scenarios.
The automatic guided vehicle is equipped with LiDAR and frontal RGB sensors and learns to reach underneath the target dolly.
NavACL-Q greatly facilitates the whole learning process and a pre-trained feature extractor manifestly boosts the training speed.
arXiv Detail & Related papers (2022-02-23T13:50:01Z) - Benchmarking Safe Deep Reinforcement Learning in Aquatic Navigation [78.17108227614928]
We propose a benchmark environment for Safe Reinforcement Learning focusing on aquatic navigation.
We consider a value-based and policy-gradient Deep Reinforcement Learning (DRL)
We also propose a verification strategy that checks the behavior of the trained models over a set of desired properties.
arXiv Detail & Related papers (2021-12-16T16:53:56Z) - Optimal control of point-to-point navigation in turbulent time-dependent
flows using Reinforcement Learning [0.0]
We present theoretical and numerical results concerning the problem to find the path that minimizes time to navigate between two given points in a complex fluid.
We show that ActorCritic algorithms are able to find quasi-optimal solutions in the presence of either time-independent or chaotically evolving flow configurations.
arXiv Detail & Related papers (2021-02-27T21:31:18Z) - Efficient Reservoir Management through Deep Reinforcement Learning [36.89242259597806]
We leverage reinforcement learning (RL) methods to compute efficient dam operation guidelines.
Specifically, we build offline simulators with real data and different mathematical models for the upstream inflow.
Experiments show that the simulator with DLM can efficiently model the inflow dynamics in the upstream and the dam operation policies trained by RL algorithms significantly outperform the human-generated policy.
arXiv Detail & Related papers (2020-12-07T16:13:05Z) - DAIS: Automatic Channel Pruning via Differentiable Annealing Indicator
Search [55.164053971213576]
convolutional neural network has achieved great success in fulfilling computer vision tasks despite large computation overhead.
Structured (channel) pruning is usually applied to reduce the model redundancy while preserving the network structure.
Existing structured pruning methods require hand-crafted rules which may lead to tremendous pruning space.
arXiv Detail & Related papers (2020-11-04T07:43:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.