Argoverse 2: Next Generation Datasets for Self-Driving Perception and
Forecasting
- URL: http://arxiv.org/abs/2301.00493v1
- Date: Mon, 2 Jan 2023 00:36:22 GMT
- Title: Argoverse 2: Next Generation Datasets for Self-Driving Perception and
Forecasting
- Authors: Benjamin Wilson, William Qi, Tanmay Agarwal, John Lambert, Jagjeet
Singh, Siddhesh Khandelwal, Bowen Pan, Ratnesh Kumar, Andrew Hartnett, Jhony
Kaesemodel Pontes, Deva Ramanan, Peter Carr, James Hays
- Abstract summary: Argoverse 2 (AV2) is a collection of three datasets for perception and forecasting research in the self-driving domain.
The Lidar dataset contains 20,000 sequences of unlabeled lidar point clouds and map-aligned pose.
The Motion Forecasting dataset contains 250,000 scenarios mined for interesting and challenging interactions between the autonomous vehicle and other actors in each local scene.
- Score: 64.7364925689825
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce Argoverse 2 (AV2) - a collection of three datasets for
perception and forecasting research in the self-driving domain. The annotated
Sensor Dataset contains 1,000 sequences of multimodal data, encompassing
high-resolution imagery from seven ring cameras, and two stereo cameras in
addition to lidar point clouds, and 6-DOF map-aligned pose. Sequences contain
3D cuboid annotations for 26 object categories, all of which are
sufficiently-sampled to support training and evaluation of 3D perception
models. The Lidar Dataset contains 20,000 sequences of unlabeled lidar point
clouds and map-aligned pose. This dataset is the largest ever collection of
lidar sensor data and supports self-supervised learning and the emerging task
of point cloud forecasting. Finally, the Motion Forecasting Dataset contains
250,000 scenarios mined for interesting and challenging interactions between
the autonomous vehicle and other actors in each local scene. Models are tasked
with the prediction of future motion for "scored actors" in each scenario and
are provided with track histories that capture object location, heading,
velocity, and category. In all three datasets, each scenario contains its own
HD Map with 3D lane and crosswalk geometry - sourced from data captured in six
distinct cities. We believe these datasets will support new and existing
machine learning research problems in ways that existing datasets do not. All
datasets are released under the CC BY-NC-SA 4.0 license.
Related papers
- Every Dataset Counts: Scaling up Monocular 3D Object Detection with Joint Datasets Training [9.272389295055271]
This study investigates the pipeline for training a monocular 3D object detection model on a diverse collection of 3D and 2D datasets.
The proposed framework comprises three components: (1) a robust monocular 3D model capable of functioning across various camera settings, (2) a selective-training strategy to accommodate datasets with differing class annotations, and (3) a pseudo 3D training approach using 2D labels to enhance detection performance in scenes containing only 2D labels.
arXiv Detail & Related papers (2023-10-02T06:17:24Z) - SUPS: A Simulated Underground Parking Scenario Dataset for Autonomous
Driving [41.221988979184665]
SUPS is a simulated dataset for underground automatic parking.
It supports multiple tasks with multiple sensors and multiple semantic labels aligned with successive images.
We also evaluate the state-of-the-art SLAM algorithms and perception models on our dataset.
arXiv Detail & Related papers (2023-02-25T02:59:12Z) - IDD-3D: Indian Driving Dataset for 3D Unstructured Road Scenes [79.18349050238413]
Preparation and training of deploy-able deep learning architectures require the models to be suited to different traffic scenarios.
An unstructured and complex driving layout found in several developing countries such as India poses a challenge to these models.
We build a new dataset, IDD-3D, which consists of multi-modal data from multiple cameras and LiDAR sensors with 12k annotated driving LiDAR frames.
arXiv Detail & Related papers (2022-10-23T23:03:17Z) - DOLPHINS: Dataset for Collaborative Perception enabled Harmonious and
Interconnected Self-driving [19.66714697653504]
Vehicle-to-Everything (V2X) network has enabled collaborative perception in autonomous driving.
The lack of datasets has severely blocked the development of collaborative perception algorithms.
We release DOLPHINS: dataset for cOllaborative Perception enabled Harmonious and INterconnected Self-driving.
arXiv Detail & Related papers (2022-07-15T17:07:07Z) - SensatUrban: Learning Semantics from Urban-Scale Photogrammetric Point
Clouds [52.624157840253204]
We introduce SensatUrban, an urban-scale UAV photogrammetry point cloud dataset consisting of nearly three billion points collected from three UK cities, covering 7.6 km2.
Each point in the dataset has been labelled with fine-grained semantic annotations, resulting in a dataset that is three times the size of the previous existing largest photogrammetric point cloud dataset.
arXiv Detail & Related papers (2022-01-12T14:48:11Z) - One Million Scenes for Autonomous Driving: ONCE Dataset [91.94189514073354]
We introduce the ONCE dataset for 3D object detection in the autonomous driving scenario.
The data is selected from 144 driving hours, which is 20x longer than the largest 3D autonomous driving dataset available.
We reproduce and evaluate a variety of self-supervised and semi-supervised methods on the ONCE dataset.
arXiv Detail & Related papers (2021-06-21T12:28:08Z) - Semantic Segmentation on Swiss3DCities: A Benchmark Study on Aerial
Photogrammetric 3D Pointcloud Dataset [67.44497676652173]
We introduce a new outdoor urban 3D pointcloud dataset, covering a total area of 2.7 $km2$, sampled from three Swiss cities.
The dataset is manually annotated for semantic segmentation with per-point labels, and is built using photogrammetry from images acquired by multirotors equipped with high-resolution cameras.
arXiv Detail & Related papers (2020-12-23T21:48:47Z) - Towards Semantic Segmentation of Urban-Scale 3D Point Clouds: A Dataset,
Benchmarks and Challenges [52.624157840253204]
We present an urban-scale photogrammetric point cloud dataset with nearly three billion richly annotated points.
Our dataset consists of large areas from three UK cities, covering about 7.6 km2 of the city landscape.
We evaluate the performance of state-of-the-art algorithms on our dataset and provide a comprehensive analysis of the results.
arXiv Detail & Related papers (2020-09-07T14:47:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.