Generalizing Decision Making for Automated Driving with an Invariant
Environment Representation using Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2102.06765v1
- Date: Fri, 12 Feb 2021 20:37:29 GMT
- Title: Generalizing Decision Making for Automated Driving with an Invariant
Environment Representation using Deep Reinforcement Learning
- Authors: Karl Kurzer, Philip Sch\"orner, Alexander Albers, Hauke Thomsen, Karam
Daaboul, J. Marius Z\"ollner
- Abstract summary: Current approaches either do not generalize well beyond the training data or are not capable to consider a variable number of traffic participants.
We propose an invariant environment representation from the perspective of the ego vehicle.
We show that the agents are capable to generalize successfully to unseen scenarios, due to the abstraction.
- Score: 55.41644538483948
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Data driven approaches for decision making applied to automated driving
require appropriate generalization strategies, to ensure applicability to the
world's variability. Current approaches either do not generalize well beyond
the training data or are not capable to consider a variable number of traffic
participants. Therefore we propose an invariant environment representation from
the perspective of the ego vehicle. The representation encodes all necessary
information for safe decision making. To assess the generalization capabilities
of the novel environment representation, we train our agents on a small subset
of scenarios and evaluate on the entire set. Here we show that the agents are
capable to generalize successfully to unseen scenarios, due to the abstraction.
In addition we present a simple occlusion model that enables our agents to
navigate intersections with occlusions without a significant change in
performance.
Related papers
- Towards Generalizable Reinforcement Learning via Causality-Guided Self-Adaptive Representations [22.6449779859417]
General intelligence requires quick adaption across tasks.
In this paper, we explore a wider range of scenarios where not only the distribution but also the environment spaces may change.
We introduce a causality-guided self-adaptive representation-based approach, called CSR, that equips the agent to generalize effectively.
arXiv Detail & Related papers (2024-07-30T08:48:49Z) - LDP: A Local Diffusion Planner for Efficient Robot Navigation and Collision Avoidance [16.81917489473445]
The conditional diffusion model has been demonstrated as an efficient tool for learning robot policies.
The intricate nature of real-world scenarios, characterized by dynamic obstacles and maze-like structures, underscores the complexity of robot local navigation decision-making.
arXiv Detail & Related papers (2024-07-02T04:53:35Z) - A Minimalist Prompt for Zero-Shot Policy Learning [61.65128628634916]
Transformer-based methods have exhibited significant generalization ability when prompted with target-domain demonstrations or example solutions during inference.
We show that conditioning a decision transformer on task parameters alone can enable zero-shot generalization on par with or better than its demonstration-conditioned counterpart.
We introduce an additional learnable prompt which is demonstrated to further boost zero-shot generalization across a range of robotic control, manipulation, and navigation benchmark tasks.
arXiv Detail & Related papers (2024-05-09T19:15:33Z) - SAFE-SIM: Safety-Critical Closed-Loop Traffic Simulation with Diffusion-Controllable Adversaries [94.84458417662407]
We introduce SAFE-SIM, a controllable closed-loop safety-critical simulation framework.
Our approach yields two distinct advantages: 1) generating realistic long-tail safety-critical scenarios that closely reflect real-world conditions, and 2) providing controllable adversarial behavior for more comprehensive and interactive evaluations.
We validate our framework empirically using the nuScenes and nuPlan datasets across multiple planners, demonstrating improvements in both realism and controllability.
arXiv Detail & Related papers (2023-12-31T04:14:43Z) - Invariance is Key to Generalization: Examining the Role of
Representation in Sim-to-Real Transfer for Visual Navigation [35.01394611106655]
Key to generalization is representations that are rich enough to capture all task-relevant information.
We experimentally study such a representation for visual navigation.
We show that our representation reduces the A-distance between the training and test domains.
arXiv Detail & Related papers (2023-10-23T15:15:19Z) - Leveraging sparse and shared feature activations for disentangled
representation learning [112.22699167017471]
We propose to leverage knowledge extracted from a diversified set of supervised tasks to learn a common disentangled representation.
We validate our approach on six real world distribution shift benchmarks, and different data modalities.
arXiv Detail & Related papers (2023-04-17T01:33:24Z) - Concept-modulated model-based offline reinforcement learning for rapid
generalization [5.512991103610139]
We propose a solution that self-generates simulated scenarios constrained by environmental concepts and dynamics learned in an unsupervised manner.
In particular, an internal model of the agent's environment is conditioned on low-dimensional concept representations of the input space that are sensitive to the agent's actions.
We show dramatic improvements in one-shot generalization to different instances of specified failure cases as well as zero-shot generalization to similar variations compared to model-based and model-free approaches.
arXiv Detail & Related papers (2022-09-07T15:06:38Z) - Dynamic Memory for Interpretable Sequential Optimisation [0.0]
We present a solution to handling non-stationarity that is suitable for deployment at scale.
We develop an adaptive Bayesian learning agent that employs a novel form of dynamic memory.
We describe the architecture of a large-scale deployment of automatic-as-a-service.
arXiv Detail & Related papers (2022-06-28T12:29:13Z) - MetaVIM: Meta Variationally Intrinsic Motivated Reinforcement Learning for Decentralized Traffic Signal Control [54.162449208797334]
Traffic signal control aims to coordinate traffic signals across intersections to improve the traffic efficiency of a district or a city.
Deep reinforcement learning (RL) has been applied to traffic signal control recently and demonstrated promising performance where each traffic signal is regarded as an agent.
We propose a novel Meta Variationally Intrinsic Motivated (MetaVIM) RL method to learn the decentralized policy for each intersection that considers neighbor information in a latent way.
arXiv Detail & Related papers (2021-01-04T03:06:08Z) - Can Autonomous Vehicles Identify, Recover From, and Adapt to
Distribution Shifts? [104.04999499189402]
Out-of-training-distribution (OOD) scenarios are a common challenge of learning agents at deployment.
We propose an uncertainty-aware planning method, called emphrobust imitative planning (RIP)
Our method can detect and recover from some distribution shifts, reducing the overconfident and catastrophic extrapolations in OOD scenes.
We introduce an autonomous car novel-scene benchmark, textttCARNOVEL, to evaluate the robustness of driving agents to a suite of tasks with distribution shifts.
arXiv Detail & Related papers (2020-06-26T11:07:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.