Falsification-Driven Reinforcement Learning for Maritime Motion Planning
- URL: http://arxiv.org/abs/2510.06970v1
- Date: Wed, 08 Oct 2025 12:56:31 GMT
- Title: Falsification-Driven Reinforcement Learning for Maritime Motion Planning
- Authors: Marlon Müller, Florian Finkeldei, Hanna Krasowski, Murat Arcak, Matthias Althoff,
- Abstract summary: Compliance with maritime traffic rules is essential for the safe operation of autonomous vessels.<n>Training reinforcement learning (RL) agents to adhere to them is challenging.<n>We propose a falsification-driven RL approach that generates adversarial training scenarios in which the vessel under test violates maritime traffic rules.
- Score: 10.405737384575334
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Compliance with maritime traffic rules is essential for the safe operation of autonomous vessels, yet training reinforcement learning (RL) agents to adhere to them is challenging. The behavior of RL agents is shaped by the training scenarios they encounter, but creating scenarios that capture the complexity of maritime navigation is non-trivial, and real-world data alone is insufficient. To address this, we propose a falsification-driven RL approach that generates adversarial training scenarios in which the vessel under test violates maritime traffic rules, which are expressed as signal temporal logic specifications. Our experiments on open-sea navigation with two vessels demonstrate that the proposed approach provides more relevant training scenarios and achieves more consistent rule compliance.
Related papers
- Discrete Diffusion for Reflective Vision-Language-Action Models in Autonomous Driving [55.13109926181247]
We introduce ReflectDrive, a learning-based framework that integrates a reflection mechanism for safe trajectory generation via discrete diffusion.<n>Central to our approach is a safety-aware reflection mechanism that performs iterative self-correction without gradient.<n>Our method begins with goal-conditioned trajectory generation to model multi-modal driving behaviors.
arXiv Detail & Related papers (2025-09-24T13:35:15Z) - FLP-XR: Future Location Prediction on Extreme Scale Maritime Data in Real-time [0.8937169040399775]
This paper introduces FLP-XR, a model that leverages maritime mobility data to construct a robust framework that offers precise predictions.<n>We demonstrate the efficiency of our approach through an extensive experimental study using three real-world AIS datasets.
arXiv Detail & Related papers (2025-03-10T13:31:42Z) - Sea-cret Agents: Maritime Abduction for Region Generation to Expose Dark Vessel Trajectories [0.6037276428689637]
Bad actors in the maritime industry engage in illegal behaviors after disabling their vessel's automatic identification system (AIS)<n>Machine learning approaches only succeed in identifying the locations of these dark vessels'' in the immediate future.<n>We combine concepts from abduction, logic programming, and rule learning to create an efficient method that approaches full recall of dark vessels.
arXiv Detail & Related papers (2025-02-03T16:36:26Z) - Towards Deviation-Robust Agent Navigation via Perturbation-Aware
Contrastive Learning [125.61772424068903]
Vision-and-language navigation (VLN) asks an agent to follow a given language instruction to navigate through a real 3D environment.
We present a model-agnostic training paradigm, called Progressive Perturbation-aware Contrastive Learning (PROPER) to enhance the generalization ability of existing VLN agents.
arXiv Detail & Related papers (2024-03-09T02:34:13Z) - Provable Traffic Rule Compliance in Safe Reinforcement Learning on the Open Sea [8.017543518311196]
Reinforcement learning (RL) is a promising method to find motion plans for autonomous vehicles.
Our approach accomplishes guaranteed rule-compliance by integrating temporal logic specifications into RL.
In numerical evaluations on critical maritime traffic situations, our agent always complies with the formalized legal rules and never collides.
arXiv Detail & Related papers (2024-02-13T14:59:19Z) - SAFE-SIM: Safety-Critical Closed-Loop Traffic Simulation with Diffusion-Controllable Adversaries [94.84458417662407]
We introduce SAFE-SIM, a controllable closed-loop safety-critical simulation framework.
Our approach yields two distinct advantages: 1) generating realistic long-tail safety-critical scenarios that closely reflect real-world conditions, and 2) providing controllable adversarial behavior for more comprehensive and interactive evaluations.
We validate our framework empirically using the nuScenes and nuPlan datasets across multiple planners, demonstrating improvements in both realism and controllability.
arXiv Detail & Related papers (2023-12-31T04:14:43Z) - Two-step dynamic obstacle avoidance [0.0]
This paper proposes a two-step architecture for handling dynamic obstacle avoidance (DOA) tasks by combining supervised and reinforcement learning (RL)
In the first step, we introduce a data-driven approach to estimate the collision risk (CR) of an obstacle using a recurrent neural network.
In the second step, we include these CR estimates into the observation space of an RL agent to increase its situational awareness.
arXiv Detail & Related papers (2023-11-28T14:55:50Z) - Learning to Sail Dynamic Networks: The MARLIN Reinforcement Learning
Framework for Congestion Control in Tactical Environments [53.08686495706487]
This paper proposes an RL framework that leverages an accurate and parallelizable emulation environment to reenact the conditions of a tactical network.
We evaluate our RL learning framework by training a MARLIN agent in conditions replicating a bottleneck link transition between a Satellite Communication (SATCOM) and an UHF Wide Band (UHF) radio link.
arXiv Detail & Related papers (2023-06-27T16:15:15Z) - Vessel-following model for inland waterways based on deep reinforcement
learning [0.0]
This study aims at investigating the feasibility of RL-based vehicle-following for complex vehicle dynamics and strong environmental disturbances.
We developed an inland waterways vessel-following model based on realistic vessel dynamics.
Our model demonstrated safe and comfortable driving in all scenarios, proving excellent generalization abilities.
arXiv Detail & Related papers (2022-07-07T12:19:03Z) - Benchmarking Safe Deep Reinforcement Learning in Aquatic Navigation [78.17108227614928]
We propose a benchmark environment for Safe Reinforcement Learning focusing on aquatic navigation.
We consider a value-based and policy-gradient Deep Reinforcement Learning (DRL)
We also propose a verification strategy that checks the behavior of the trained models over a set of desired properties.
arXiv Detail & Related papers (2021-12-16T16:53:56Z) - Deep Structured Reactive Planning [94.92994828905984]
We propose a novel data-driven, reactive planning objective for self-driving vehicles.
We show that our model outperforms a non-reactive variant in successfully completing highly complex maneuvers.
arXiv Detail & Related papers (2021-01-18T01:43:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.