DRUformer: Enhancing the driving scene Important object detection with
driving relationship self-understanding
- URL: http://arxiv.org/abs/2311.06497v2
- Date: Thu, 14 Dec 2023 04:08:48 GMT
- Title: DRUformer: Enhancing the driving scene Important object detection with
driving relationship self-understanding
- Authors: Yingjie Niu, Ming Ding, Keisuke Fujii, Kento Ohtani, Alexander
Carballo, Kazuya Takeda
- Abstract summary: Traffic accidents frequently lead to fatal injuries, contributing to over 50 million deaths until 2023.
Previous research primarily assessed the importance of individual participants, treating them as independent entities.
We introduce Driving scene Relationship self-Understanding transformer (DRUformer) to enhance the important object detection task.
- Score: 50.81809690183755
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Traffic accidents frequently lead to fatal injuries, contributing to over 50
million deaths until 2023. To mitigate driving hazards and ensure personal
safety, it is crucial to assist vehicles in anticipating important objects
during travel. Previous research on important object detection primarily
assessed the importance of individual participants, treating them as
independent entities and frequently overlooking the connections between these
participants. Unfortunately, this approach has proven less effective in
detecting important objects in complex scenarios. In response, we introduce
Driving scene Relationship self-Understanding transformer (DRUformer), designed
to enhance the important object detection task. The DRUformer is a
transformer-based multi-modal important object detection model that takes into
account the relationships between all the participants in the driving scenario.
Recognizing that driving intention also significantly affects the detection of
important objects during driving, we have incorporated a module for embedding
driving intention. To assess the performance of our approach, we conducted a
comparative experiment on the DRAMA dataset, pitting our model against other
state-of-the-art (SOTA) models. The results demonstrated a noteworthy 16.2\%
improvement in mIoU and a substantial 12.3\% boost in ACC compared to SOTA
methods. Furthermore, we conducted a qualitative analysis of our model's
ability to detect important objects across different road scenarios and
classes, highlighting its effectiveness in diverse contexts. Finally, we
conducted various ablation studies to assess the efficiency of the proposed
modules in our DRUformer model.
Related papers
- On-Road Object Importance Estimation: A New Dataset and A Model with Multi-Fold Top-Down Guidance [70.80612792049315]
This paper contributes a new large-scale dataset named Traffic Object Importance (TOI)
It proposes a model that integrates multi-fold top-down guidance with the bottom-up feature.
Our model outperforms state-of-the-art methods by large margins.
arXiv Detail & Related papers (2024-11-26T06:37:10Z) - Effects of Real-Life Traffic Sign Alteration on YOLOv7- an Object
Recognition Model [1.6334452280183571]
This study investigates the influence of altered traffic signs on the accuracy and effectiveness of object recognition.
It employs a publicly available dataset to introduce alterations in shape, color, content, visibility, angles and background.
The study demonstrates a notable decline in detection and classification accuracy when confronted with traffic signs in unusual conditions.
arXiv Detail & Related papers (2023-05-09T14:51:29Z) - Camera-Radar Perception for Autonomous Vehicles and ADAS: Concepts,
Datasets and Metrics [77.34726150561087]
This work aims to carry out a study on the current scenario of camera and radar-based perception for ADAS and autonomous vehicles.
Concepts and characteristics related to both sensors, as well as to their fusion, are presented.
We give an overview of the Deep Learning-based detection and segmentation tasks, and the main datasets, metrics, challenges, and open questions in vehicle perception.
arXiv Detail & Related papers (2023-03-08T00:48:32Z) - Suspected Object Matters: Rethinking Model's Prediction for One-stage
Visual Grounding [93.82542533426766]
We propose a Suspected Object Transformation mechanism (SOT) to encourage the target object selection among the suspected ones.
SOT can be seamlessly integrated into existing CNN and Transformer-based one-stage visual grounders.
Extensive experiments demonstrate the effectiveness of our proposed method.
arXiv Detail & Related papers (2022-03-10T06:41:07Z) - Important Object Identification with Semi-Supervised Learning for
Autonomous Driving [37.654878298744855]
We propose a novel approach for important object identification in egocentric driving scenarios.
We present a semi-supervised learning pipeline to enable the model to learn from unlimited unlabeled data.
Our approach also outperforms rule-based baselines by a large margin.
arXiv Detail & Related papers (2022-03-05T01:23:13Z) - Evaluating Object (mis)Detection from a Safety and Reliability
Perspective: Discussion and Measures [1.8492669447784602]
We propose new object detection measures that reward the correct identification of objects that are most dangerous and most likely to affect driving decisions.
We apply our model on the recent autonomous driving dataset nuScenes, and we compare nine object detectors.
Results show that, in several settings, object detectors that perform best according to the nuScenes ranking are not the preferable ones when the focus is shifted on safety and reliability.
arXiv Detail & Related papers (2022-03-04T09:31:20Z) - Modeling Interactions of Autonomous Vehicles and Pedestrians with Deep
Multi-Agent Reinforcement Learning for Collision Avoidance [20.542143534865154]
In this work, we model the corresponding interaction sequence as a Markov decision process (MDP) that is solved by deep reinforcement learning (DRL) algorithms.
The presented PCAM system with different levels of intelligent pedestrian behavior is benchmarked according to the agents' collision rate and the resulting traffic flow efficiency.
The results show that the AV is able to completely mitigate collisions under the majority of the investigated conditions.
arXiv Detail & Related papers (2021-09-30T17:06:39Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - Studying Person-Specific Pointing and Gaze Behavior for Multimodal
Referencing of Outside Objects from a Moving Vehicle [58.720142291102135]
Hand pointing and eye gaze have been extensively investigated in automotive applications for object selection and referencing.
Existing outside-the-vehicle referencing methods focus on a static situation, whereas the situation in a moving vehicle is highly dynamic and subject to safety-critical constraints.
We investigate the specific characteristics of each modality and the interaction between them when used in the task of referencing outside objects.
arXiv Detail & Related papers (2020-09-23T14:56:19Z) - Interaction Graphs for Object Importance Estimation in On-road Driving
Videos [9.344790309080283]
Learning to estimate the importance of each object on the driver's real-time decision-making may help better understand human driving behavior.
We propose a novel framework for object importance estimation using an interaction graph.
arXiv Detail & Related papers (2020-03-12T22:28:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.