WOMD-LiDAR: Raw Sensor Dataset Benchmark for Motion Forecasting
- URL: http://arxiv.org/abs/2304.03834v2
- Date: Sun, 18 Feb 2024 23:23:01 GMT
- Title: WOMD-LiDAR: Raw Sensor Dataset Benchmark for Motion Forecasting
- Authors: Kan Chen, Runzhou Ge, Hang Qiu, Rami AI-Rfou, Charles R. Qi, Xuanyu
Zhou, Zoey Yang, Scott Ettinger, Pei Sun, Zhaoqi Leng, Mustafa Baniodeh, Ivan
Bogun, Weiyue Wang, Mingxing Tan, Dragomir Anguelov
- Abstract summary: We augment the Open Motion dataset with large-scale, high-quality, diverse LiDAR data for the motion forecasting task.
The new augmented dataset WOMD-LiDAR consists of over 100,000 scenes that each spans 20 seconds, consisting of well-synchronized and high quality LiDAR point clouds captured across a range of urban and suburban geographies.
Experiments show that the LiDAR data brings improvement in the motion forecasting task.
- Score: 38.95768804529958
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Widely adopted motion forecasting datasets substitute the observed sensory
inputs with higher-level abstractions such as 3D boxes and polylines. These
sparse shapes are inferred through annotating the original scenes with
perception systems' predictions. Such intermediate representations tie the
quality of the motion forecasting models to the performance of computer vision
models. Moreover, the human-designed explicit interfaces between perception and
motion forecasting typically pass only a subset of the semantic information
present in the original sensory input. To study the effect of these modular
approaches, design new paradigms that mitigate these limitations, and
accelerate the development of end-to-end motion forecasting models, we augment
the Waymo Open Motion Dataset (WOMD) with large-scale, high-quality, diverse
LiDAR data for the motion forecasting task.
The new augmented dataset WOMD-LiDAR consists of over 100,000 scenes that
each spans 20 seconds, consisting of well-synchronized and calibrated high
quality LiDAR point clouds captured across a range of urban and suburban
geographies (https://waymo.com/open/data/motion/). Compared to Waymo Open
Dataset (WOD), WOMD-LiDAR dataset contains 100x more scenes. Furthermore, we
integrate the LiDAR data into the motion forecasting model training and provide
a strong baseline. Experiments show that the LiDAR data brings improvement in
the motion forecasting task. We hope that WOMD-LiDAR will provide new
opportunities for boosting end-to-end motion forecasting models.
Related papers
- Future Does Matter: Boosting 3D Object Detection with Temporal Motion Estimation in Point Cloud Sequences [25.74000325019015]
We introduce a novel LiDAR 3D object detection framework, namely LiSTM, to facilitate spatial-temporal feature learning with cross-frame motion forecasting information.
We have conducted experiments on the aggregation and nuScenes datasets to demonstrate that the proposed framework achieves superior 3D detection performance.
arXiv Detail & Related papers (2024-09-06T16:29:04Z) - LiveHPS: LiDAR-based Scene-level Human Pose and Shape Estimation in Free
Environment [59.320414108383055]
We present LiveHPS, a novel single-LiDAR-based approach for scene-level human pose and shape estimation.
We propose a huge human motion dataset, named FreeMotion, which is collected in various scenarios with diverse human poses.
arXiv Detail & Related papers (2024-02-27T03:08:44Z) - Realistic Rainy Weather Simulation for LiDARs in CARLA Simulator [17.605152926277707]
We propose a simulator-based physical modeling approach to augment LiDAR data in rainy weather.
We complete the modeling task of the rainy weather in the CARLA simulator and establish a pipeline for LiDAR data collection.
In the experiment, we observe that the model augmented by the synthetic data improves the object detection task's performance.
arXiv Detail & Related papers (2023-12-20T05:16:04Z) - JRDB-Traj: A Dataset and Benchmark for Trajectory Forecasting in Crowds [79.00975648564483]
Trajectory forecasting models, employed in fields such as robotics, autonomous vehicles, and navigation, face challenges in real-world scenarios.
This dataset provides comprehensive data, including the locations of all agents, scene images, and point clouds, all from the robot's perspective.
The objective is to predict the future positions of agents relative to the robot using raw sensory input data.
arXiv Detail & Related papers (2023-11-05T18:59:31Z) - LiDAR Data Synthesis with Denoising Diffusion Probabilistic Models [1.1965844936801797]
Generative modeling of 3D LiDAR data is an emerging task with promising applications for autonomous mobile robots.
We present R2DM, a novel generative model for LiDAR data that can generate diverse and high-fidelity 3D scene point clouds.
Our method is built upon denoising diffusion probabilistic models (DDPMs), which have shown impressive results among generative model frameworks.
arXiv Detail & Related papers (2023-09-17T12:26:57Z) - Learning to Simulate Realistic LiDARs [66.7519667383175]
We introduce a pipeline for data-driven simulation of a realistic LiDAR sensor.
We show that our model can learn to encode realistic effects such as dropped points on transparent surfaces.
We use our technique to learn models of two distinct LiDAR sensors and use them to improve simulated LiDAR data accordingly.
arXiv Detail & Related papers (2022-09-22T13:12:54Z) - LiDARCap: Long-range Marker-less 3D Human Motion Capture with LiDAR
Point Clouds [58.402752909624716]
Existing motion capture datasets are largely short-range and cannot yet fit the need of long-range applications.
We propose LiDARHuman26M, a new human motion capture dataset captured by LiDAR at a much longer range to overcome this limitation.
Our dataset also includes the ground truth human motions acquired by the IMU system and the synchronous RGB images.
arXiv Detail & Related papers (2022-03-28T12:52:45Z) - Scalable Scene Flow from Point Clouds in the Real World [30.437100097997245]
We introduce a new large scale benchmark for scene flow based on the Open dataset.
We show how previous works were bounded based on the amount of real LiDAR data available.
We introduce the model architecture FastFlow3D that provides real time inference on the full point cloud.
arXiv Detail & Related papers (2021-03-01T20:56:05Z) - SLPC: a VRNN-based approach for stochastic lidar prediction and
completion in autonomous driving [63.87272273293804]
We propose a new LiDAR prediction framework that is based on generative models namely Variational Recurrent Neural Networks (VRNNs)
Our algorithm is able to address the limitations of previous video prediction frameworks when dealing with sparse data by spatially inpainting the depth maps in the upcoming frames.
We present a sparse version of VRNNs and an effective self-supervised training method that does not require any labels.
arXiv Detail & Related papers (2021-02-19T11:56:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.