Salient Sign Detection In Safe Autonomous Driving: AI Which Reasons Over
Full Visual Context
- URL: http://arxiv.org/abs/2301.05804v1
- Date: Sat, 14 Jan 2023 01:47:09 GMT
- Title: Salient Sign Detection In Safe Autonomous Driving: AI Which Reasons Over
Full Visual Context
- Authors: Ross Greer, Akshay Gopalkrishnan, Nachiket Deo, Akshay Rangesh, Mohan
Trivedi
- Abstract summary: Various traffic signs in a driving scene have an unequal impact on the driver's decisions.
We construct a traffic sign detection model which emphasizes performance on salient signs.
We show that a model trained with Salience-Sensitive Focal Loss outperforms a model trained without.
- Score: 2.799896314754614
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Detecting road traffic signs and accurately determining how they can affect
the driver's future actions is a critical task for safe autonomous driving
systems. However, various traffic signs in a driving scene have an unequal
impact on the driver's decisions, making detecting the salient traffic signs a
more important task. Our research addresses this issue, constructing a traffic
sign detection model which emphasizes performance on salient signs, or signs
that influence the decisions of a driver. We define a traffic sign salience
property and use it to construct the LAVA Salient Signs Dataset, the first
traffic sign dataset that includes an annotated salience property. Next, we use
a custom salience loss function, Salience-Sensitive Focal Loss, to train a
Deformable DETR object detection model in order to emphasize stronger
performance on salient signs. Results show that a model trained with
Salience-Sensitive Focal Loss outperforms a model trained without, with regards
to recall of both salient signs and all signs combined. Further, the
performance margin on salient signs compared to all signs is largest for the
model trained with Salience-Sensitive Focal Loss.
Related papers
- DriveCoT: Integrating Chain-of-Thought Reasoning with End-to-End Driving [81.04174379726251]
This paper collects a comprehensive end-to-end driving dataset named DriveCoT.
It contains sensor data, control decisions, and chain-of-thought labels to indicate the reasoning process.
We propose a baseline model called DriveCoT-Agent, trained on our dataset, to generate chain-of-thought predictions and final decisions.
arXiv Detail & Related papers (2024-03-25T17:59:01Z) - DRUformer: Enhancing the driving scene Important object detection with
driving relationship self-understanding [50.81809690183755]
Traffic accidents frequently lead to fatal injuries, contributing to over 50 million deaths until 2023.
Previous research primarily assessed the importance of individual participants, treating them as independent entities.
We introduce Driving scene Relationship self-Understanding transformer (DRUformer) to enhance the important object detection task.
arXiv Detail & Related papers (2023-11-11T07:26:47Z) - Adversary ML Resilience in Autonomous Driving Through Human Centered
Perception Mechanisms [0.0]
This paper explores the resilience of autonomous driving systems against three main physical adversarial attacks (tape, graffiti, illumination)
To build robustness against attacks, defense techniques like adversarial training and transfer learning were implemented.
Results demonstrated transfer learning models played a crucial role in performance by allowing knowledge gained from shape training to improve generalizability of road sign classification.
arXiv Detail & Related papers (2023-11-02T04:11:45Z) - Robust Traffic Light Detection Using Salience-Sensitive Loss:
Computational Framework and Evaluations [0.3061098887924466]
This paper proposes a traffic light detection model which focuses on defining salient lights as the lights that affect the driver's future decisions.
We then use this salience property to construct the LAVA Salient Lights dataset, the first US traffic light dataset with an annotated salience property.
We train a Deformable DETR object detection transformer model using Salience-Sensitive Focal Loss to emphasize stronger performance on salient traffic lights.
arXiv Detail & Related papers (2023-05-08T07:22:15Z) - OpenLane-V2: A Topology Reasoning Benchmark for Unified 3D HD Mapping [84.65114565766596]
We present OpenLane-V2, the first dataset on topology reasoning for traffic scene structure.
OpenLane-V2 consists of 2,000 annotated road scenes that describe traffic elements and their correlation to the lanes.
We evaluate various state-of-the-art methods, and present their quantitative and qualitative results on OpenLane-V2 to indicate future avenues for investigating topology reasoning in traffic scenes.
arXiv Detail & Related papers (2023-04-20T16:31:22Z) - CueCAn: Cue Driven Contextual Attention For Identifying Missing Traffic
Signs on Unconstrained Roads [26.649617412538717]
Missing or non-existing object detection has been studied for locating missing curbs and estimating reasonable regions for pedestrians on road scene images.
We present the first and most challenging video dataset for missing objects, with multiple types of traffic signs for which the cues are visible without the signs in the scenes.
We train the encoder to classify the presence of traffic sign cues and then train the entire segmentation model end-to-end to localize missing traffic signs.
arXiv Detail & Related papers (2023-03-05T11:06:20Z) - On Salience-Sensitive Sign Classification in Autonomous Vehicle Path
Planning: Experimental Explorations with a Novel Dataset [11.007092387379076]
We present a dataset with a novel feature, sign salience, defined to indicate whether a sign is distinctly informative to the goals of the ego vehicle.
Using convolutional networks on cropped signs, in tandem with experimental augmentation by road type, image coordinates, and planned maneuver, we predict the sign salience property with 76% accuracy.
arXiv Detail & Related papers (2021-12-02T02:45:30Z) - Automated Object Behavioral Feature Extraction for Potential Risk
Analysis based on Video Sensor [6.291501119156943]
Pedestrians are exposed to risk of death or serious injuries on roads, especially unsignalized crosswalks.
We propose an automated and simpler system for effectively extracting object behavioral features from video sensors deployed on the road.
This study demonstrates the potential for a network of connected video sensors to provide actionable data for smart cities.
arXiv Detail & Related papers (2021-07-08T01:11:31Z) - Driving-Signal Aware Full-Body Avatars [49.89791440532946]
We present a learning-based method for building driving-signal aware full-body avatars.
Our model is a conditional variational autoencoder that can be animated with incomplete driving signals.
We demonstrate the efficacy of our approach on the challenging problem of full-body animation for virtual telepresence.
arXiv Detail & Related papers (2021-05-21T16:22:38Z) - Detecting 32 Pedestrian Attributes for Autonomous Vehicles [103.87351701138554]
In this paper, we address the problem of jointly detecting pedestrians and recognizing 32 pedestrian attributes.
We introduce a Multi-Task Learning (MTL) model relying on a composite field framework, which achieves both goals in an efficient way.
We show competitive detection and attribute recognition results, as well as a more stable MTL training.
arXiv Detail & Related papers (2020-12-04T15:10:12Z) - Driver Intention Anticipation Based on In-Cabin and Driving Scene
Monitoring [52.557003792696484]
We present a framework for the detection of the drivers' intention based on both in-cabin and traffic scene videos.
Our framework achieves a prediction with the accuracy of 83.98% and F1-score of 84.3%.
arXiv Detail & Related papers (2020-06-20T11:56:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.