Adversarial Mixture Density Networks: Learning to Drive Safely from
Collision Data
- URL: http://arxiv.org/abs/2107.04485v1
- Date: Fri, 9 Jul 2021 15:16:30 GMT
- Title: Adversarial Mixture Density Networks: Learning to Drive Safely from
Collision Data
- Authors: Sampo Kuutti, Saber Fallah, Richard Bowden
- Abstract summary: Imitation learning has been widely used to learn control policies for autonomous driving based on pre-recorded data.
We introduce Adversarial Mixture Density Networks (AMDN), which learns two distributions from separate datasets.
We show that AMDN provides significant benefits for the safety of the learned control policy, when compared to pure imitation learning or standard mixture density network approaches.
- Score: 42.57240271305088
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Imitation learning has been widely used to learn control policies for
autonomous driving based on pre-recorded data. However, imitation learning
based policies have been shown to be susceptible to compounding errors when
encountering states outside of the training distribution. Further, these agents
have been demonstrated to be easily exploitable by adversarial road users
aiming to create collisions. To overcome these shortcomings, we introduce
Adversarial Mixture Density Networks (AMDN), which learns two distributions
from separate datasets. The first is a distribution of safe actions learned
from a dataset of naturalistic human driving. The second is a distribution
representing unsafe actions likely to lead to collision, learned from a dataset
of collisions. During training, we leverage these two distributions to provide
an additional loss based on the similarity of the two distributions. By
penalising the safe action distribution based on its similarity to the unsafe
action distribution when training on the collision dataset, a more robust and
safe control policy is obtained. We demonstrate the proposed AMDN approach in a
vehicle following use-case, and evaluate under naturalistic and adversarial
testing environments. We show that despite its simplicity, AMDN provides
significant benefits for the safety of the learned control policy, when
compared to pure imitation learning or standard mixture density network
approaches.
Related papers
- Closing the Distribution Gap in Adversarial Training for LLMs [50.33186122381395]
Adversarial training for LLMs is one of the most promising methods to reliably improve robustness against adversaries.<n>We argue that current adversarial training algorithms minimize adversarial loss on their training set but inadequately cover the data distribution, resulting in vulnerability to seemingly simple attacks.<n>We propose Distributional Adversarial Training, DAT, to approximate the true joint distribution of prompts and responses.
arXiv Detail & Related papers (2026-02-16T22:34:52Z) - Learning collision risk proactively from naturalistic driving data at scale [3.1457219084519004]
This study introduces the Generalised Surrogate Safety Measure (GSSM)<n>GSSM learns collision risk from naturalistic driving without the need for crash or risk labels.<n> Diverse data from naturalistic driving, including motion kinematics, weather, lighting, etc., are used to train multiple GSSMs.<n>A basic GSSM using only instantaneous motion kinematics achieves an area under the precision-recall curve of 0.9 and secures a median time advance of 2.6 seconds to prevent potential collisions.
arXiv Detail & Related papers (2025-05-19T07:22:32Z) - Safety Monitoring for Learning-Enabled Cyber-Physical Systems in Out-of-Distribution Scenarios [17.629563106665557]
We propose to directly monitor safety in a manner that is itself robust to OOD data.
Our safety monitor additionally uses a novel combination of adaptive conformal prediction and incremental learning.
arXiv Detail & Related papers (2025-04-18T05:42:37Z) - Enhancing Autonomous Driving Safety with Collision Scenario Integration [36.83682052117178]
We propose SafeFusion, a training framework to learn from collision data.
Instead of over-relying on imitation learning, SafeFusion integrates safety-oriented metrics during training to enable collision avoidance learning.
We also propose CollisionGen, a scalable data generation pipeline to generate diverse, high-quality scenarios.
arXiv Detail & Related papers (2025-03-05T23:08:43Z) - FuzzRisk: Online Collision Risk Estimation for Autonomous Vehicles based on Depth-Aware Object Detection via Fuzzy Inference [6.856508678236828]
The framework takes two sets of predictions from different algorithms and associates their inconsistencies with the collision risk via fuzzy inference.
We experimentally validate that, based on Intersection-over-Union (IoU) and a depth discrepancy measure, the inconsistencies between the two sets of predictions strongly correlate to the error of the 3D object detector.
We optimize the fuzzy inference system towards an existing offline metric that matches AV collision rates well.
arXiv Detail & Related papers (2024-11-09T20:20:36Z) - Collision Probability Distribution Estimation via Temporal Difference Learning [0.46085106405479537]
We introduce CollisionPro, a pioneering framework designed to estimate cumulative collision probability distributions.
We formulate our framework within the context of reinforcement learning to pave the way for safety-aware agents.
A comprehensive examination of our framework is conducted using a realistic autonomous driving simulator.
arXiv Detail & Related papers (2024-07-29T13:32:42Z) - Federated Adversarial Learning for Robust Autonomous Landing Runway Detection [6.029462194041386]
In this paper, we propose a federated adversarial learning-based framework to detect landing runways.
To the best of our knowledge, this marks the first instance of federated learning work that address the adversarial sample problem in landing runway detection.
arXiv Detail & Related papers (2024-06-22T19:31:52Z) - Defending against Data Poisoning Attacks in Federated Learning via User Elimination [0.0]
This paper introduces a novel framework focused on the strategic elimination of adversarial users within a federated model.
We detect anomalies in the aggregation phase of the Federated Algorithm, by integrating metadata gathered by the local training instances with Differential Privacy techniques.
Our experiments demonstrate the efficacy of our methods, significantly mitigating the risk of data poisoning while maintaining user privacy and model performance.
arXiv Detail & Related papers (2024-04-19T10:36:00Z) - FedDriveScore: Federated Scoring Driving Behavior with a Mixture of
Metric Distributions [6.195950768412144]
Vehicle-cloud collaboration is proposed as a privacy-friendly alternative to centralized learning.
This framework includes a consistently federated version of the scoring method to reduce the performance degradation of the global scoring model.
arXiv Detail & Related papers (2024-01-13T02:15:41Z) - DALA: A Distribution-Aware LoRA-Based Adversarial Attack against
Language Models [64.79319733514266]
Adversarial attacks can introduce subtle perturbations to input data.
Recent attack methods can achieve a relatively high attack success rate (ASR)
We propose a Distribution-Aware LoRA-based Adversarial Attack (DALA) method.
arXiv Detail & Related papers (2023-11-14T23:43:47Z) - Semi-supervised Domain Adaptive Structure Learning [72.01544419893628]
Semi-supervised domain adaptation (SSDA) is a challenging problem requiring methods to overcome both 1) overfitting towards poorly annotated data and 2) distribution shift across domains.
We introduce an adaptive structure learning method to regularize the cooperation of SSL and DA.
arXiv Detail & Related papers (2021-12-12T06:11:16Z) - Adversarial Robustness through the Lens of Causality [105.51753064807014]
adversarial vulnerability of deep neural networks has attracted significant attention in machine learning.
We propose to incorporate causality into mitigating adversarial vulnerability.
Our method can be seen as the first attempt to leverage causality for mitigating adversarial vulnerability.
arXiv Detail & Related papers (2021-06-11T06:55:02Z) - Privacy-Preserving Federated Learning on Partitioned Attributes [6.661716208346423]
Federated learning empowers collaborative training without exposing local data or models.
We introduce an adversarial learning based procedure which tunes a local model to release privacy-preserving intermediate representations.
To alleviate the accuracy decline, we propose a defense method based on the forward-backward splitting algorithm.
arXiv Detail & Related papers (2021-04-29T14:49:14Z) - Learning while Respecting Privacy and Robustness to Distributional
Uncertainties and Adversarial Data [66.78671826743884]
The distributionally robust optimization framework is considered for training a parametric model.
The objective is to endow the trained model with robustness against adversarially manipulated input data.
Proposed algorithms offer robustness with little overhead.
arXiv Detail & Related papers (2020-07-07T18:25:25Z) - Adversarial Self-Supervised Contrastive Learning [62.17538130778111]
Existing adversarial learning approaches mostly use class labels to generate adversarial samples that lead to incorrect predictions.
We propose a novel adversarial attack for unlabeled data, which makes the model confuse the instance-level identities of the perturbed data samples.
We present a self-supervised contrastive learning framework to adversarially train a robust neural network without labeled data.
arXiv Detail & Related papers (2020-06-13T08:24:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.