Receding Moving Object Segmentation in 3D LiDAR Data Using Sparse 4D
Convolutions
- URL: http://arxiv.org/abs/2206.04129v1
- Date: Wed, 8 Jun 2022 18:51:14 GMT
- Title: Receding Moving Object Segmentation in 3D LiDAR Data Using Sparse 4D
Convolutions
- Authors: Benedikt Mersch, Xieyuanli Chen, Ignacio Vizzo, Lucas Nunes, Jens
Behley, Cyrill Stachniss
- Abstract summary: We tackle the problem of distinguishing 3D LiDAR points that belong to currently moving objects, like walking pedestrians or driving cars, from points that are obtained from non-moving objects, like walls but also parked cars.
Our approach takes a sequence of observed LiDAR scans and turns them into a voxelized sparse 4D point cloud.
We apply computationally efficient sparse 4D convolutions to jointly extract spatial and temporal features and predict moving object confidence scores for all points in the sequence.
- Score: 33.538055872850514
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A key challenge for autonomous vehicles is to navigate in unseen dynamic
environments. Separating moving objects from static ones is essential for
navigation, pose estimation, and understanding how other traffic participants
are likely to move in the near future. In this work, we tackle the problem of
distinguishing 3D LiDAR points that belong to currently moving objects, like
walking pedestrians or driving cars, from points that are obtained from
non-moving objects, like walls but also parked cars. Our approach takes a
sequence of observed LiDAR scans and turns them into a voxelized sparse 4D
point cloud. We apply computationally efficient sparse 4D convolutions to
jointly extract spatial and temporal features and predict moving object
confidence scores for all points in the sequence. We develop a receding horizon
strategy that allows us to predict moving objects online and to refine
predictions on the go based on new observations. We use a binary Bayes filter
to recursively integrate new predictions of a scan resulting in more robust
estimation. We evaluate our approach on the SemanticKITTI moving object
segmentation challenge and show more accurate predictions than existing
methods. Since our approach only operates on the geometric information of point
clouds over time, it generalizes well to new, unseen environments, which we
evaluate on the Apollo dataset.
Related papers
- SeMoLi: What Moves Together Belongs Together [51.72754014130369]
We tackle semi-supervised object detection based on motion cues.
Recent results suggest that motion-based clustering methods can be used to pseudo-label instances of moving objects.
We re-think this approach and suggest that both, object detection, as well as motion-inspired pseudo-labeling, can be tackled in a data-driven manner.
arXiv Detail & Related papers (2024-02-29T18:54:53Z) - DORT: Modeling Dynamic Objects in Recurrent for Multi-Camera 3D Object
Detection and Tracking [67.34803048690428]
We propose to model Dynamic Objects in RecurrenT (DORT) to tackle this problem.
DORT extracts object-wise local volumes for motion estimation that also alleviates the heavy computational burden.
It is flexible and practical that can be plugged into most camera-based 3D object detectors.
arXiv Detail & Related papers (2023-03-29T12:33:55Z) - InsMOS: Instance-Aware Moving Object Segmentation in LiDAR Data [13.196031553445117]
We propose a novel network that addresses the challenge of segmenting moving objects in 3D LiDAR scans.
Our method exploits a sequence of point clouds as input and quantifies them into 4D voxels.
We use 4D sparse convolutions to extract motion features from the 4D voxels and inject them into the current scan.
arXiv Detail & Related papers (2023-03-07T14:12:52Z) - Point Cloud Forecasting as a Proxy for 4D Occupancy Forecasting [58.45661235893729]
One promising self-supervised task is 3D point cloud forecasting from unannotated LiDAR sequences.
We show that this task requires algorithms to implicitly capture (1) sensor extrinsics (i.e., the egomotion of the autonomous vehicle), (2) sensor intrinsics (i.e., the sampling pattern specific to the particular LiDAR sensor), and (3) the shape and motion of other objects in the scene.
We render point cloud data from 4D occupancy predictions given sensor extrinsics and intrinsics, allowing one to train and test occupancy algorithms with unannotated LiDAR sequences.
arXiv Detail & Related papers (2023-02-25T18:12:37Z) - Lifting 2D Object Locations to 3D by Discounting LiDAR Outliers across
Objects and Views [70.1586005070678]
We present a system for automatically converting 2D mask object predictions and raw LiDAR point clouds into full 3D bounding boxes of objects.
Our method significantly outperforms previous work despite the fact that those methods use significantly more complex pipelines, 3D models and additional human-annotated external sources of prior information.
arXiv Detail & Related papers (2021-09-16T13:01:13Z) - Monocular Quasi-Dense 3D Object Tracking [99.51683944057191]
A reliable and accurate 3D tracking framework is essential for predicting future locations of surrounding objects and planning the observer's actions in numerous applications such as autonomous driving.
We propose a framework that can effectively associate moving objects over time and estimate their full 3D bounding box information from a sequence of 2D images captured on a moving platform.
arXiv Detail & Related papers (2021-03-12T15:30:02Z) - Auto4D: Learning to Label 4D Objects from Sequential Point Clouds [89.30951657004408]
We propose an automatic pipeline that generates accurate object trajectories in 3D space from LiDAR point clouds.
The key idea is to decompose the 4D object label into two parts: the object size in 3D that's fixed through time for rigid objects, and the motion path describing the evolution of the object's pose through time.
Given the cheap but noisy input, our model produces higher quality 4D labels by re-estimating the object size and smoothing the motion path.
arXiv Detail & Related papers (2021-01-17T04:23:05Z) - Tracking from Patterns: Learning Corresponding Patterns in Point Clouds
for 3D Object Tracking [34.40019455462043]
We propose to learn 3D object correspondences from temporal point cloud data and infer the motion information from correspondence patterns.
Our method exceeds the existing 3D tracking methods on both the KITTI and larger scale Nuscenes dataset.
arXiv Detail & Related papers (2020-10-20T06:07:20Z) - Any Motion Detector: Learning Class-agnostic Scene Dynamics from a
Sequence of LiDAR Point Clouds [4.640835690336654]
We propose a novel real-time approach of temporal context aggregation for motion detection and motion parameters estimation.
We introduce an ego-motion compensation layer to achieve real-time inference with performance comparable to a naive odometric transform of the original point cloud sequence.
arXiv Detail & Related papers (2020-04-24T10:40:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.