Generalized dynamic cognitive hierarchy models for strategic driving
behavior
- URL: http://arxiv.org/abs/2109.09861v1
- Date: Mon, 20 Sep 2021 21:49:52 GMT
- Title: Generalized dynamic cognitive hierarchy models for strategic driving
behavior
- Authors: Atrisha Sarkar, Kate Larson, Krzysztof Czarnecki
- Abstract summary: We develop a framework of generalized dynamic cognitive hierarchy for both modelling naturalistic human driving behavior and behavior planning for autonomous vehicles.
Based on evaluation on two large naturalistic datasets, we show that automata strategies are well suited for level-0 behavior in a dynamic level-k framework.
- Score: 13.415452801139843
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While there has been an increasing focus on the use of game theoretic models
for autonomous driving, empirical evidence shows that there are still open
questions around dealing with the challenges of common knowledge assumptions as
well as modeling bounded rationality. To address some of these practical
challenges, we develop a framework of generalized dynamic cognitive hierarchy
for both modelling naturalistic human driving behavior as well as behavior
planning for autonomous vehicles (AV). This framework is built upon a rich
model of level-0 behavior through the use of automata strategies, an
interpretable notion of bounded rationality through safety and maneuver
satisficing, and a robust response for planning. Based on evaluation on two
large naturalistic datasets as well as simulation of critical traffic
scenarios, we show that i) automata strategies are well suited for level-0
behavior in a dynamic level-k framework, and ii) the proposed robust response
to a heterogeneous population of strategic and non-strategic reasoners can be
an effective approach for game theoretic planning in AV.
Related papers
- Reinforcement Learning under Latent Dynamics: Toward Statistical and Algorithmic Modularity [51.40558987254471]
Real-world applications of reinforcement learning often involve environments where agents operate on complex, high-dimensional observations.
This paper addresses the question of reinforcement learning under $textitgeneral$ latent dynamics from a statistical and algorithmic perspective.
arXiv Detail & Related papers (2024-10-23T14:22:49Z) - Adversarial Safety-Critical Scenario Generation using Naturalistic Human Driving Priors [2.773055342671194]
We introduce a natural adversarial scenario generation solution using naturalistic human driving priors and reinforcement learning techniques.
Our findings demonstrate that the proposed model can generate realistic safety-critical test scenarios covering both naturalness and adversariality.
arXiv Detail & Related papers (2024-08-06T13:58:56Z) - Learning to Steer Markovian Agents under Model Uncertainty [23.603487812521657]
We study how to design additional rewards to steer multi-agent systems towards desired policies.
Motivated by the limitation of existing works, we consider a new category of learning dynamics called emphMarkovian agents
We learn a emphhistory-dependent steering strategy to handle the inherent model uncertainty about the agents' learning dynamics.
arXiv Detail & Related papers (2024-07-14T14:01:38Z) - SAFE-SIM: Safety-Critical Closed-Loop Traffic Simulation with Diffusion-Controllable Adversaries [94.84458417662407]
We introduce SAFE-SIM, a controllable closed-loop safety-critical simulation framework.
Our approach yields two distinct advantages: 1) generating realistic long-tail safety-critical scenarios that closely reflect real-world conditions, and 2) providing controllable adversarial behavior for more comprehensive and interactive evaluations.
We validate our framework empirically using the nuScenes and nuPlan datasets across multiple planners, demonstrating improvements in both realism and controllability.
arXiv Detail & Related papers (2023-12-31T04:14:43Z) - Dual policy as self-model for planning [71.73710074424511]
We refer to the model used to simulate one's decisions as the agent's self-model.
Inspired by current reinforcement learning approaches and neuroscience, we explore the benefits and limitations of using a distilled policy network as the self-model.
arXiv Detail & Related papers (2023-06-07T13:58:45Z) - Hierarchical Model-Based Imitation Learning for Planning in Autonomous
Driving [47.59287162318435]
We demonstrate the first large-scale application of model-based generative adversarial imitation learning (MGAIL) to the task of dense urban self-driving.
We augment standard MGAIL using a hierarchical model to enable generalization to arbitrary goal routes, and measure performance using a closed-loop evaluation framework with simulated interactive agents.
We train policies from expert trajectories collected from real vehicles driving over 100,000 miles in San Francisco, and demonstrate a steerable policy that can navigate robustly even in a zero-shot setting.
arXiv Detail & Related papers (2022-10-18T02:15:34Z) - UMBRELLA: Uncertainty-Aware Model-Based Offline Reinforcement Learning
Leveraging Planning [1.1339580074756188]
Offline reinforcement learning (RL) provides a framework for learning decision-making from offline data.
Self-driving vehicles (SDV) learn a policy, which potentially even outperforms the behavior in the sub-optimal data set.
This motivates the use of model-based offline RL approaches, which leverage planning.
arXiv Detail & Related papers (2021-11-22T10:37:52Z) - A taxonomy of strategic human interactions in traffic conflicts [13.415452801139843]
We develop a taxonomy for strategic interactions based on patterns of interaction in traffic conflicts.
We demonstrate a process of automatic mapping of strategies generated by a strategic planner to the categories in the taxonomy.
We evaluate two popular solution concepts used in strategic planning in AVs, QLk and Subgame perfect $epsilon$-Nash Equilibrium.
arXiv Detail & Related papers (2021-09-27T22:19:30Z) - Solution Concepts in Hierarchical Games under Bounded Rationality with Applications to Autonomous Driving [8.500525426182115]
We create game theoretic models of driving behaviour using hierarchical games.
We evaluate the behaviour models on the basis of model fit to naturalistic data, as well as their predictive capacity.
Our results suggest that among the behaviour models evaluated, at the level of maneuvers, modeling driving behaviour as an adaptation of the Quantal level-k model with level-0 behaviour modelled as pure rule-following provides the best fit to naturalistic driving behaviour.
arXiv Detail & Related papers (2020-09-21T17:13:50Z) - Goal-Aware Prediction: Learning to Model What Matters [105.43098326577434]
One of the fundamental challenges in using a learned forward dynamics model is the mismatch between the objective of the learned model and that of the downstream planner or policy.
We propose to direct prediction towards task relevant information, enabling the model to be aware of the current task and encouraging it to only model relevant quantities of the state space.
We find that our method more effectively models the relevant parts of the scene conditioned on the goal, and as a result outperforms standard task-agnostic dynamics models and model-free reinforcement learning.
arXiv Detail & Related papers (2020-07-14T16:42:59Z) - Context-aware Dynamics Model for Generalization in Model-Based
Reinforcement Learning [124.9856253431878]
We decompose the task of learning a global dynamics model into two stages: (a) learning a context latent vector that captures the local dynamics, then (b) predicting the next state conditioned on it.
In order to encode dynamics-specific information into the context latent vector, we introduce a novel loss function that encourages the context latent vector to be useful for predicting both forward and backward dynamics.
The proposed method achieves superior generalization ability across various simulated robotics and control tasks, compared to existing RL schemes.
arXiv Detail & Related papers (2020-05-14T08:10:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.