3D Object Detection with a Self-supervised Lidar Scene Flow Backbone
- URL: http://arxiv.org/abs/2205.00705v1
- Date: Mon, 2 May 2022 07:53:29 GMT
- Title: 3D Object Detection with a Self-supervised Lidar Scene Flow Backbone
- Authors: Eme\c{c} Er\c{c}elik, Ekim Yurtsever, Mingyu Liu, Zhijie Yang, Hanzhen
Zhang, P{\i}nar Top\c{c}am, Maximilian Listl, Y{\i}lmaz Kaan \c{C}ayl{\i},
Alois Knoll
- Abstract summary: We propose using a self-supervised training strategy to learn a general point cloud backbone model for downstream 3D vision tasks.
Our main contribution leverages learned flow and motion representations and combines a self-supervised backbone with a 3D detection head.
Experiments on KITTI and nuScenes benchmarks show that the proposed self-supervised pre-training increases 3D detection performance significantly.
- Score: 10.341296683155973
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: State-of-the-art 3D detection methods rely on supervised learning and large
labelled datasets. However, annotating lidar data is resource-consuming, and
depending only on supervised learning limits the applicability of trained
models. Against this backdrop, here we propose using a self-supervised training
strategy to learn a general point cloud backbone model for downstream 3D vision
tasks. 3D scene flow can be estimated with self-supervised learning using cycle
consistency, which removes labelled data requirements. Moreover, the perception
of objects in the traffic scenarios heavily relies on making sense of the
sparse data in the spatio-temporal context. Our main contribution leverages
learned flow and motion representations and combines a self-supervised backbone
with a 3D detection head focusing mainly on the relation between the scene flow
and detection tasks. In this way, self-supervised scene flow training
constructs point motion features in the backbone, which help distinguish
objects based on their different motion patterns used with a 3D detection head.
Experiments on KITTI and nuScenes benchmarks show that the proposed
self-supervised pre-training increases 3D detection performance significantly.
Related papers
- Enhancing Generalizability of Representation Learning for Data-Efficient 3D Scene Understanding [50.448520056844885]
We propose a generative Bayesian network to produce diverse synthetic scenes with real-world patterns.
A series of experiments robustly display our method's consistent superiority over existing state-of-the-art pre-training approaches.
arXiv Detail & Related papers (2024-06-17T07:43:53Z) - AGO-Net: Association-Guided 3D Point Cloud Object Detection Network [86.10213302724085]
We propose a novel 3D detection framework that associates intact features for objects via domain adaptation.
We achieve new state-of-the-art performance on the KITTI 3D detection benchmark in both accuracy and speed.
arXiv Detail & Related papers (2022-08-24T16:54:38Z) - Time-to-Label: Temporal Consistency for Self-Supervised Monocular 3D
Object Detection [46.077668660248534]
We argue that the temporal consistency on the level of object poses, provides an important supervision signal.
Specifically, we propose a self-supervised loss which uses this consistency, in addition to render-and-compare losses.
We finetune a synthetically trained monocular 3D object detection model using the pseudo-labels that we generated on real data.
arXiv Detail & Related papers (2022-03-04T08:55:49Z) - The Devil is in the Task: Exploiting Reciprocal Appearance-Localization
Features for Monocular 3D Object Detection [62.1185839286255]
Low-cost monocular 3D object detection plays a fundamental role in autonomous driving.
We introduce a Dynamic Feature Reflecting Network, named DFR-Net.
We rank 1st among all the monocular 3D object detectors in the KITTI test set.
arXiv Detail & Related papers (2021-12-28T07:31:18Z) - Spatio-temporal Self-Supervised Representation Learning for 3D Point
Clouds [96.9027094562957]
We introduce a-temporal representation learning framework, capable of learning from unlabeled tasks.
Inspired by how infants learn from visual data in the wild, we explore rich cues derived from the 3D data.
STRL takes two temporally-related frames from a 3D point cloud sequence as the input, transforms it with the spatial data augmentation, and learns the invariant representation self-supervisedly.
arXiv Detail & Related papers (2021-09-01T04:17:11Z) - Occlusion Guided Self-supervised Scene Flow Estimation on 3D Point
Clouds [4.518012967046983]
Understanding the flow in 3D space of sparsely sampled points between two consecutive time frames is the core stone of modern geometric-driven systems.
This work presents a new self-supervised training method and an architecture for the 3D scene flow estimation under occlusions.
arXiv Detail & Related papers (2021-04-10T09:55:19Z) - Monocular Quasi-Dense 3D Object Tracking [99.51683944057191]
A reliable and accurate 3D tracking framework is essential for predicting future locations of surrounding objects and planning the observer's actions in numerous applications such as autonomous driving.
We propose a framework that can effectively associate moving objects over time and estimate their full 3D bounding box information from a sequence of 2D images captured on a moving platform.
arXiv Detail & Related papers (2021-03-12T15:30:02Z) - Do not trust the neighbors! Adversarial Metric Learning for
Self-Supervised Scene Flow Estimation [0.0]
Scene flow is the task of estimating 3D motion vectors to individual points of a dynamic 3D scene.
We propose a 3D scene flow benchmark and a novel self-supervised setup for training flow models.
We find that our setup is able to keep motion coherence and preserve local geometries, which many self-supervised baselines fail to grasp.
arXiv Detail & Related papers (2020-11-01T17:41:32Z) - PointContrast: Unsupervised Pre-training for 3D Point Cloud
Understanding [107.02479689909164]
In this work, we aim at facilitating research on 3D representation learning.
We measure the effect of unsupervised pre-training on a large source set of 3D scenes.
arXiv Detail & Related papers (2020-07-21T17:59:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.