Learning-based Preference Prediction for Constrained Multi-Criteria
Path-Planning
- URL: http://arxiv.org/abs/2108.01080v1
- Date: Mon, 2 Aug 2021 17:13:45 GMT
- Title: Learning-based Preference Prediction for Constrained Multi-Criteria
Path-Planning
- Authors: Kevin Osanlou, Christophe Guettier, Andrei Bursuc, Tristan Cazenave
and Eric Jacopin
- Abstract summary: Constrained path-planning for Autonomous Ground Vehicles (AGV) is one such application.
We leverage knowledge acquired through offline simulations by training a neural network model to predict the uncertain criterion.
We integrate this model inside a path-planner which can solve problems online.
- Score: 12.457788665461312
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Learning-based methods are increasingly popular for search algorithms in
single-criterion optimization problems. In contrast, for multiple-criteria
optimization there are significantly fewer approaches despite the existence of
numerous applications. Constrained path-planning for Autonomous Ground Vehicles
(AGV) is one such application, where an AGV is typically deployed in disaster
relief or search and rescue applications in off-road environments. The agent
can be faced with the following dilemma : optimize a source-destination path
according to a known criterion and an uncertain criterion under operational
constraints. The known criterion is associated to the cost of the path,
representing the distance. The uncertain criterion represents the feasibility
of driving through the path without requiring human intervention. It depends on
various external parameters such as the physics of the vehicle, the state of
the explored terrains or weather conditions. In this work, we leverage
knowledge acquired through offline simulations by training a neural network
model to predict the uncertain criterion. We integrate this model inside a
path-planner which can solve problems online. Finally, we conduct experiments
on realistic AGV scenarios which illustrate that the proposed framework
requires human intervention less frequently, trading for a limited increase in
the path distance.
Related papers
- Model-based RL as a Minimalist Approach to Horizon-Free and Second-Order Bounds [59.875550175217874]
We show that a simple Model-based Reinforcement Learning scheme achieves strong regret and sample bounds in online and offline RL settings.
We highlight that our algorithms are simple, fairly standard, and indeed have been extensively studied in the RL literature.
arXiv Detail & Related papers (2024-08-16T19:52:53Z) - Learning-Aided Warmstart of Model Predictive Control in Uncertain
Fast-Changing Traffic [2.0965639599405366]
We use a network based multimodal predictor to generate proposals for the autonomous vehicle trajectory.
This approach enables us to identify multiple local minima and provide an improved initial guess.
We validate our approach with Monte Carlo simulations distinct scenarios.
arXiv Detail & Related papers (2023-10-04T16:00:21Z) - Integrating Higher-Order Dynamics and Roadway-Compliance into
Constrained ILQR-based Trajectory Planning for Autonomous Vehicles [3.200238632208686]
Trajectory planning aims to produce a globally optimal route for Autonomous Passenger Vehicles.
Existing implementations utilizing the vehicle bicycle kinematic model may not guarantee controllable trajectories.
We augment this model by higher-order terms, including the first and second-order derivatives of curvature and longitudinal jerk.
arXiv Detail & Related papers (2023-09-25T22:30:18Z) - Dual Formulation for Chance Constrained Stochastic Shortest Path with
Application to Autonomous Vehicle Behavior Planning [3.655021726150368]
The Constrained Shortest Path problem (C-SSP) is a formalism for planning in environments under certain types of operating constraints.
This work's first contribution is an exact integer linear formulation for Chance-constrained policies.
Third, we show that the CC-SSP formalism can be generalized to account for constraints that span through multiple time steps.
arXiv Detail & Related papers (2023-02-25T16:40:00Z) - Interactively Learning Preference Constraints in Linear Bandits [100.78514640066565]
We study sequential decision-making with known rewards and unknown constraints.
As an application, we consider learning constraints to represent human preferences in a driving simulation.
arXiv Detail & Related papers (2022-06-10T17:52:58Z) - Offline Stochastic Shortest Path: Learning, Evaluation and Towards
Optimality [57.91411772725183]
In this paper, we consider the offline shortest path problem when the state space and the action space are finite.
We design the simple value-based algorithms for tackling both offline policy evaluation (OPE) and offline policy learning tasks.
Our analysis of these simple algorithms yields strong instance-dependent bounds which can imply worst-case bounds that are near-minimax optimal.
arXiv Detail & Related papers (2022-06-10T07:44:56Z) - Motion Planning for Autonomous Vehicles in the Presence of Uncertainty
Using Reinforcement Learning [0.0]
Motion planning under uncertainty is one of the main challenges in developing autonomous driving vehicles.
We propose a reinforcement learning based solution to manage uncertainty by optimizing for the worst case outcome.
The proposed approach yields much better motion planning behavior compared to conventional RL algorithms and behaves comparably to humans driving style.
arXiv Detail & Related papers (2021-10-01T20:32:25Z) - Optimal Solving of Constrained Path-Planning Problems with Graph
Convolutional Networks and Optimized Tree Search [12.457788665461312]
We propose a hybrid solving planner that combines machine learning models and an optimal solver.
We conduct experiments on realistic scenarios and show that GCN support enables substantial speedup and smoother scaling to harder problems.
arXiv Detail & Related papers (2021-08-02T16:53:21Z) - Congestion-aware Multi-agent Trajectory Prediction for Collision
Avoidance [110.63037190641414]
We propose to learn congestion patterns explicitly and devise a novel "Sense--Learn--Reason--Predict" framework.
By decomposing the learning phases into two stages, a "student" can learn contextual cues from a "teacher" while generating collision-free trajectories.
In experiments, we demonstrate that the proposed model is able to generate collision-free trajectory predictions in a synthetic dataset.
arXiv Detail & Related papers (2021-03-26T02:42:33Z) - Reinforcement Learning for Low-Thrust Trajectory Design of
Interplanetary Missions [77.34726150561087]
This paper investigates the use of reinforcement learning for the robust design of interplanetary trajectories in presence of severe disturbances.
An open-source implementation of the state-of-the-art algorithm Proximal Policy Optimization is adopted.
The resulting Guidance and Control Network provides both a robust nominal trajectory and the associated closed-loop guidance law.
arXiv Detail & Related papers (2020-08-19T15:22:15Z) - Can Autonomous Vehicles Identify, Recover From, and Adapt to
Distribution Shifts? [104.04999499189402]
Out-of-training-distribution (OOD) scenarios are a common challenge of learning agents at deployment.
We propose an uncertainty-aware planning method, called emphrobust imitative planning (RIP)
Our method can detect and recover from some distribution shifts, reducing the overconfident and catastrophic extrapolations in OOD scenes.
We introduce an autonomous car novel-scene benchmark, textttCARNOVEL, to evaluate the robustness of driving agents to a suite of tasks with distribution shifts.
arXiv Detail & Related papers (2020-06-26T11:07:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.