Where and What: Driver Attention-based Object Detection
- URL: http://arxiv.org/abs/2204.12150v1
- Date: Tue, 26 Apr 2022 08:38:22 GMT
- Title: Where and What: Driver Attention-based Object Detection
- Authors: Yao Rong, Naemi-Rebecca Kassautzki, Wolfgang Fuhl, Enkelejda Kasneci
- Abstract summary: We bridge the gap between pixel-level and object-level attention prediction.
Our framework achieves competitive state-of-the-art performance on both pixel-level and object-level.
- Score: 13.5947650184579
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Human drivers use their attentional mechanisms to focus on critical objects
and make decisions while driving. As human attention can be revealed from gaze
data, capturing and analyzing gaze information has emerged in recent years to
benefit autonomous driving technology. Previous works in this context have
primarily aimed at predicting "where" human drivers look at and lack knowledge
of "what" objects drivers focus on. Our work bridges the gap between
pixel-level and object-level attention prediction. Specifically, we propose to
integrate an attention prediction module into a pretrained object detection
framework and predict the attention in a grid-based style. Furthermore,
critical objects are recognized based on predicted attended-to areas. We
evaluate our proposed method on two driver attention datasets, BDD-A and
DR(eye)VE. Our framework achieves competitive state-of-the-art performance in
the attention prediction on both pixel-level and object-level but is far more
efficient (75.3 GFLOPs less) in computation.
Related papers
- OOSTraj: Out-of-Sight Trajectory Prediction With Vision-Positioning Denoising [49.86409475232849]
Trajectory prediction is fundamental in computer vision and autonomous driving.
Existing approaches in this field often assume precise and complete observational data.
We present a novel method for out-of-sight trajectory prediction that leverages a vision-positioning technique.
arXiv Detail & Related papers (2024-04-02T18:30:29Z) - Implicit Occupancy Flow Fields for Perception and Prediction in
Self-Driving [68.95178518732965]
A self-driving vehicle (SDV) must be able to perceive its surroundings and predict the future behavior of other traffic participants.
Existing works either perform object detection followed by trajectory of the detected objects, or predict dense occupancy and flow grids for the whole scene.
This motivates our unified approach to perception and future prediction that implicitly represents occupancy and flow over time with a single neural network.
arXiv Detail & Related papers (2023-08-02T23:39:24Z) - Visual Saliency Detection in Advanced Driver Assistance Systems [7.455416595124159]
We present an intelligent system that combines a drowsiness detection system for drivers with a scene comprehension pipeline based on saliency.
We employ an innovative biosensor embedded on the car steering wheel to monitor the driver.
A dedicated 1D temporal deep convolutional network has been devised to classify the collected PPG time-series.
arXiv Detail & Related papers (2023-07-26T15:41:54Z) - Pedestrian Crossing Action Recognition and Trajectory Prediction with 3D
Human Keypoints [25.550524178542833]
We propose a novel multi-task learning framework for pedestrian crossing action recognition and trajectory prediction.
We use 3D human keypoints extracted from raw sensor data to capture rich information on human pose and activity.
We show that our approach achieves state-of-the-art performance on a wide range of evaluation metrics.
arXiv Detail & Related papers (2023-06-01T18:27:48Z) - CoCAtt: A Cognitive-Conditioned Driver Attention Dataset (Supplementary
Material) [31.888206001447625]
Driver attention prediction can play an instrumental role in mitigating and preventing high-risk events.
We present a new driver attention dataset, CoCAtt.
CoCAtt is the largest and the most diverse driver attention dataset in terms of autonomy levels, eye tracker resolutions, and driving scenarios.
arXiv Detail & Related papers (2022-07-08T17:35:17Z) - 3D Object Detection for Autonomous Driving: A Comprehensive Survey [48.30753402458884]
3D object detection, which intelligently predicts the locations, sizes, and categories of the critical 3D objects near an autonomous vehicle, is an important part of a perception system.
This paper reviews the advances in 3D object detection for autonomous driving.
arXiv Detail & Related papers (2022-06-19T19:43:11Z) - CoCAtt: A Cognitive-Conditioned Driver Attention Dataset [16.177399201198636]
Driver attention prediction can play an instrumental role in mitigating and preventing high-risk events.
We present a new driver attention dataset, CoCAtt.
CoCAtt is the largest and the most diverse driver attention dataset in terms of autonomy levels, eye tracker resolutions, and driving scenarios.
arXiv Detail & Related papers (2021-11-19T02:42:34Z) - TRiPOD: Human Trajectory and Pose Dynamics Forecasting in the Wild [77.59069361196404]
TRiPOD is a novel method for predicting body dynamics based on graph attentional networks.
To incorporate a real-world challenge, we learn an indicator representing whether an estimated body joint is visible/invisible at each frame.
Our evaluation shows that TRiPOD outperforms all prior work and state-of-the-art specifically designed for each of the trajectory and pose forecasting tasks.
arXiv Detail & Related papers (2021-04-08T20:01:00Z) - Detecting Invisible People [58.49425715635312]
We re-purpose tracking benchmarks and propose new metrics for the task of detecting invisible objects.
We demonstrate that current detection and tracking systems perform dramatically worse on this task.
Second, we build dynamic models that explicitly reason in 3D, making use of observations produced by state-of-the-art monocular depth estimation networks.
arXiv Detail & Related papers (2020-12-15T16:54:45Z) - Perceive, Attend, and Drive: Learning Spatial Attention for Safe
Self-Driving [84.59201486239908]
We propose an end-to-end self-driving network featuring a sparse attention module that learns to automatically attend to important regions of the input.
The attention module specifically targets motion planning, whereas prior literature only applied attention in perception tasks.
arXiv Detail & Related papers (2020-11-02T17:47:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.