Bi-Level Optimization Augmented with Conditional Variational Autoencoder
for Autonomous Driving in Dense Traffic
- URL: http://arxiv.org/abs/2212.02224v1
- Date: Mon, 5 Dec 2022 12:56:42 GMT
- Title: Bi-Level Optimization Augmented with Conditional Variational Autoencoder
for Autonomous Driving in Dense Traffic
- Authors: Arun Kumar Singh, Jatan Shrestha, Nicola Albarella
- Abstract summary: This paper presents a parameterized bi-level optimization that jointly computes the optimal behavioural decisions and the resulting trajectory.
Our approach runs in real-time using a custom GPU-accelerated batch, and a Variational Autoencoder learnt warm-start strategy.
Our approach outperforms state-of-the-art model predictive control and RL approaches in terms of collision rate while being competitive in driving efficiency.
- Score: 0.9281671380673306
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Autonomous driving has a natural bi-level structure. The goal of the upper
behavioural layer is to provide appropriate lane change, speeding up, and
braking decisions to optimize a given driving task. However, this layer can
only indirectly influence the driving efficiency through the lower-level
trajectory planner, which takes in the behavioural inputs to produce motion
commands. Existing sampling-based approaches do not fully exploit the strong
coupling between the behavioural and planning layer. On the other hand,
end-to-end Reinforcement Learning (RL) can learn a behavioural layer while
incorporating feedback from the lower-level planner. However, purely
data-driven approaches often fail in safety metrics in unseen environments.
This paper presents a novel alternative; a parameterized bi-level optimization
that jointly computes the optimal behavioural decisions and the resulting
downstream trajectory. Our approach runs in real-time using a custom
GPU-accelerated batch optimizer, and a Conditional Variational Autoencoder
learnt warm-start strategy. Extensive simulations show that our approach
outperforms state-of-the-art model predictive control and RL approaches in
terms of collision rate while being competitive in driving efficiency.
Related papers
- Understanding Optimization in Deep Learning with Central Flows [53.66160508990508]
We show that an RMS's implicit behavior can be explicitly captured by a "central flow:" a differential equation.
We show that these flows can empirically predict long-term optimization trajectories of generic neural networks.
arXiv Detail & Related papers (2024-10-31T17:58:13Z) - End-to-end Driving in High-Interaction Traffic Scenarios with Reinforcement Learning [24.578178308010912]
We propose an end-to-end model-based RL algorithm named Ramble to address these issues.
By learning a dynamics model of the environment, Ramble can foresee upcoming traffic events and make more informed, strategic decisions.
Ramble achieves state-of-the-art performance regarding route completion rate and driving score on the CARLA Leaderboard 2.0, showcasing its effectiveness in managing complex and dynamic traffic situations.
arXiv Detail & Related papers (2024-10-03T06:45:59Z) - DiFSD: Ego-Centric Fully Sparse Paradigm with Uncertainty Denoising and Iterative Refinement for Efficient End-to-End Autonomous Driving [55.53171248839489]
We propose an ego-centric fully sparse paradigm, named DiFSD, for end-to-end self-driving.
Specifically, DiFSD mainly consists of sparse perception, hierarchical interaction and iterative motion planner.
Experiments conducted on nuScenes dataset demonstrate the superior planning performance and great efficiency of DiFSD.
arXiv Detail & Related papers (2024-09-15T15:55:24Z) - Autonomous Vehicle Controllers From End-to-End Differentiable Simulation [60.05963742334746]
We propose a differentiable simulator and design an analytic policy gradients (APG) approach to training AV controllers.
Our proposed framework brings the differentiable simulator into an end-to-end training loop, where gradients of environment dynamics serve as a useful prior to help the agent learn a more grounded policy.
We find significant improvements in performance and robustness to noise in the dynamics, as well as overall more intuitive human-like handling.
arXiv Detail & Related papers (2024-09-12T11:50:06Z) - MetaFollower: Adaptable Personalized Autonomous Car Following [63.90050686330677]
We propose an adaptable personalized car-following framework - MetaFollower.
We first utilize Model-Agnostic Meta-Learning (MAML) to extract common driving knowledge from various CF events.
We additionally combine Long Short-Term Memory (LSTM) and Intelligent Driver Model (IDM) to reflect temporal heterogeneity with high interpretability.
arXiv Detail & Related papers (2024-06-23T15:30:40Z) - AutoFT: Learning an Objective for Robust Fine-Tuning [60.641186718253735]
Foundation models encode rich representations that can be adapted to downstream tasks by fine-tuning.
Current approaches to robust fine-tuning use hand-crafted regularization techniques.
We propose AutoFT, a data-driven approach for robust fine-tuning.
arXiv Detail & Related papers (2024-01-18T18:58:49Z) - Integrating Higher-Order Dynamics and Roadway-Compliance into
Constrained ILQR-based Trajectory Planning for Autonomous Vehicles [3.200238632208686]
Trajectory planning aims to produce a globally optimal route for Autonomous Passenger Vehicles.
Existing implementations utilizing the vehicle bicycle kinematic model may not guarantee controllable trajectories.
We augment this model by higher-order terms, including the first and second-order derivatives of curvature and longitudinal jerk.
arXiv Detail & Related papers (2023-09-25T22:30:18Z) - A Safe Hierarchical Planning Framework for Complex Driving Scenarios
based on Reinforcement Learning [23.007323699176467]
We propose a hierarchical behavior planning framework with a set of low-level safe controllers and a high-level reinforcement learning algorithm (H-CtRL) as a coordinator for the low-level controllers.
Safety is guaranteed by the low-level optimization/sampling-based controllers, while the high-level reinforcement learning algorithm makes H-CtRL an adaptive and efficient behavior planner.
The proposed H-CtRL is proved to be effective in various realistic simulation scenarios, with satisfying performance in terms of both safety and efficiency.
arXiv Detail & Related papers (2021-01-17T20:45:42Z) - Amortized Q-learning with Model-based Action Proposals for Autonomous
Driving on Highways [10.687104237121408]
We introduce a Reinforcement Learning based approach that coupled with a trajectory planner, learns an optimal long-term driving strategy.
By online generating locally optimal maneuvers as actions, we balance between the infinite low-level continuous action space and the limited flexibility of a fixed number of predefined standard lane-change actions.
arXiv Detail & Related papers (2020-12-06T11:04:40Z) - Efficient Sampling-Based Maximum Entropy Inverse Reinforcement Learning
with Application to Autonomous Driving [35.44498286245894]
We present an efficient sampling-based maximum-entropy inverse reinforcement learning (IRL) algorithm in this paper.
We evaluate the proposed algorithm on real driving data, including both non-interactive and interactive scenarios.
arXiv Detail & Related papers (2020-06-22T01:41:13Z) - Guided Constrained Policy Optimization for Dynamic Quadrupedal Robot
Locomotion [78.46388769788405]
We introduce guided constrained policy optimization (GCPO), an RL framework based upon our implementation of constrained policy optimization (CPPO)
We show that guided constrained RL offers faster convergence close to the desired optimum resulting in an optimal, yet physically feasible, robotic control behavior without the need for precise reward function tuning.
arXiv Detail & Related papers (2020-02-22T10:15:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.