ROAD: The ROad event Awareness Dataset for Autonomous Driving
- URL: http://arxiv.org/abs/2102.11585v2
- Date: Thu, 25 Feb 2021 10:07:31 GMT
- Title: ROAD: The ROad event Awareness Dataset for Autonomous Driving
- Authors: Gurkirt Singh, Stephen Akrigg, Manuele Di Maio, Valentina Fontana,
Reza Javanmard Alitappeh, Suman Saha, Kossar Jeddisaravi, Farzad Yousefi,
Jacob Culley, Tom Nicholson, Jordan Omokeowa, Salman Khan, Stanislao
Grazioso, Andrew Bradley, Giuseppe Di Gironimo, Fabio Cuzzolin
- Abstract summary: ROAD is designed to test an autonomous vehicle's ability to detect road events.
It comprises 22 videos, annotated with bounding boxes showing the location in the image plane of each road event.
We also provide as baseline a new incremental algorithm for online road event awareness, based on RetinaNet along time.
- Score: 16.24547478826027
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Humans approach driving in a holistic fashion which entails, in particular,
understanding road events and their evolution. Injecting these capabilities in
an autonomous vehicle has thus the potential to take situational awareness and
decision making closer to human-level performance. To this purpose, we
introduce the ROad event Awareness Dataset (ROAD) for Autonomous Driving, to
our knowledge the first of its kind. ROAD is designed to test an autonomous
vehicle's ability to detect road events, defined as triplets composed by a
moving agent, the action(s) it performs and the corresponding scene locations.
ROAD comprises 22 videos, originally from the Oxford RobotCar Dataset,
annotated with bounding boxes showing the location in the image plane of each
road event. We also provide as baseline a new incremental algorithm for online
road event awareness, based on inflating RetinaNet along time, which achieves a
mean average precision of 16.8% and 6.1% for frame-level and video-level event
detection, respectively, at 50% overlap. Though promising, these figures
highlight the challenges faced by situation awareness in autonomous driving.
Finally, ROAD allows scholars to investigate exciting tasks such as complex
(road) activity detection, future road event anticipation and the modelling of
sentient road agents in terms of mental states. Dataset can be obtained from
https://github.com/gurkirt/road-dataset and baseline code from
https://github.com/gurkirt/3D-RetinaNet.
Related papers
- ROAD-Waymo: Action Awareness at Scale for Autonomous Driving [17.531603453254434]
ROAD-Waymo is an extensive dataset for the development and benchmarking of techniques for agent, action, location and event detection in road scenes.
Considerably larger and more challenging than any existing dataset (and encompassing multiple cities), it comes with 198k annotated video frames, 54k agent tubes, 3.9M bounding boxes and a total of 12.4M labels.
arXiv Detail & Related papers (2024-11-03T20:46:50Z) - RSRD: A Road Surface Reconstruction Dataset and Benchmark for Safe and
Comfortable Autonomous Driving [67.09546127265034]
Road surface reconstruction helps to enhance the analysis and prediction of vehicle responses for motion planning and control systems.
We introduce the Road Surface Reconstruction dataset, a real-world, high-resolution, and high-precision dataset collected with a specialized platform in diverse driving conditions.
It covers common road types containing approximately 16,000 pairs of stereo images, original point clouds, and ground-truth depth/disparity maps.
arXiv Detail & Related papers (2023-10-03T17:59:32Z) - DeepAccident: A Motion and Accident Prediction Benchmark for V2X
Autonomous Driving [76.29141888408265]
We propose a large-scale dataset containing diverse accident scenarios that frequently occur in real-world driving.
The proposed DeepAccident dataset includes 57K annotated frames and 285K annotated samples, approximately 7 times more than the large-scale nuScenes dataset.
arXiv Detail & Related papers (2023-04-03T17:37:00Z) - Exploring Contextual Representation and Multi-Modality for End-to-End
Autonomous Driving [58.879758550901364]
Recent perception systems enhance spatial understanding with sensor fusion but often lack full environmental context.
We introduce a framework that integrates three cameras to emulate the human field of view, coupled with top-down bird-eye-view semantic data to enhance contextual representation.
Our method achieves displacement error by 0.67m in open-loop settings, surpassing current methods by 6.9% on the nuScenes dataset.
arXiv Detail & Related papers (2022-10-13T05:56:20Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - Detecting 32 Pedestrian Attributes for Autonomous Vehicles [103.87351701138554]
In this paper, we address the problem of jointly detecting pedestrians and recognizing 32 pedestrian attributes.
We introduce a Multi-Task Learning (MTL) model relying on a composite field framework, which achieves both goals in an efficient way.
We show competitive detection and attribute recognition results, as well as a more stable MTL training.
arXiv Detail & Related papers (2020-12-04T15:10:12Z) - VRUNet: Multi-Task Learning Model for Intent Prediction of Vulnerable
Road Users [3.6265173818019947]
We propose a multi-task learning model to predict pedestrian actions, crossing intent and forecast their future path from video sequences.
We have trained the model on naturalistic driving open-source JAAD dataset, which is rich in behavioral annotations and real world scenarios.
Experimental results show state-of-the-art performance on JAAD dataset.
arXiv Detail & Related papers (2020-07-10T14:02:25Z) - VehicleNet: Learning Robust Visual Representation for Vehicle
Re-identification [116.1587709521173]
We propose to build a large-scale vehicle dataset (called VehicleNet) by harnessing four public vehicle datasets.
We design a simple yet effective two-stage progressive approach to learning more robust visual representation from VehicleNet.
We achieve the state-of-art accuracy of 86.07% mAP on the private test set of AICity Challenge.
arXiv Detail & Related papers (2020-04-14T05:06:38Z) - The Pedestrian Patterns Dataset [11.193504036335503]
The dataset was collected by repeatedly traversing the same three routes for one week starting at different specific timeslots.
The purpose of the dataset is to capture the patterns of social and pedestrian behavior along the traversed routes at different times.
arXiv Detail & Related papers (2020-01-06T23:58:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.