FAST3D: Flow-Aware Self-Training for 3D Object Detectors
- URL: http://arxiv.org/abs/2110.09355v1
- Date: Mon, 18 Oct 2021 14:32:05 GMT
- Title: FAST3D: Flow-Aware Self-Training for 3D Object Detectors
- Authors: Christian Fruhwirth-Reisinger, Michael Opitz, Horst Possegger, Horst
Bischof
- Abstract summary: State-of-the-art self-training approaches mostly ignore the temporal nature of autonomous driving data.
We propose a flow-aware self-training method that enables unsupervised domain adaptation for 3D object detectors on continuous LiDAR point clouds.
Our results show a significant improvement over the state-of-the-art, without any prior target domain knowledge.
- Score: 12.511087244102036
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In the field of autonomous driving, self-training is widely applied to
mitigate distribution shifts in LiDAR-based 3D object detectors. This
eliminates the need for expensive, high-quality labels whenever the environment
changes (e.g., geographic location, sensor setup, weather condition).
State-of-the-art self-training approaches, however, mostly ignore the temporal
nature of autonomous driving data. To address this issue, we propose a
flow-aware self-training method that enables unsupervised domain adaptation for
3D object detectors on continuous LiDAR point clouds. In order to get reliable
pseudo-labels, we leverage scene flow to propagate detections through time. In
particular, we introduce a flow-based multi-target tracker, that exploits flow
consistency to filter and refine resulting tracks. The emerged precise
pseudo-labels then serve as a basis for model re-training. Starting with a
pre-trained KITTI model, we conduct experiments on the challenging Waymo Open
Dataset to demonstrate the effectiveness of our approach. Without any prior
target domain knowledge, our results show a significant improvement over the
state-of-the-art.
Related papers
- Finetuning Pre-trained Model with Limited Data for LiDAR-based 3D Object Detection by Bridging Domain Gaps [8.897884780881535]
LiDAR-based 3D object detectors often fail to adapt well to target domains with different sensor configurations.
Recent studies suggest that pre-trained backbones can be learned in a self-supervised manner with large-scale unlabeled LiDAR frames.
We propose a novel method, called Domain Adaptive Distill-Tuning (DADT), to adapt a pre-trained model with limited target data.
arXiv Detail & Related papers (2024-10-02T08:22:42Z) - LISO: Lidar-only Self-Supervised 3D Object Detection [25.420879730860936]
We introduce a novel self-supervised method to train SOTA lidar object detection networks.
It works on unlabeled sequences of lidar point clouds only.
It utilizes a SOTA self-supervised lidar scene flow network under the hood to generate, track, and iteratively refine pseudo ground truth.
arXiv Detail & Related papers (2024-03-11T18:02:52Z) - Unsupervised Domain Adaptation for Self-Driving from Past Traversal
Features [69.47588461101925]
We propose a method to adapt 3D object detectors to new driving environments.
Our approach enhances LiDAR-based detection models using spatial quantized historical features.
Experiments on real-world datasets demonstrate significant improvements.
arXiv Detail & Related papers (2023-09-21T15:00:31Z) - Unsupervised Adaptation from Repeated Traversals for Autonomous Driving [54.59577283226982]
Self-driving cars must generalize to the end-user's environment to operate reliably.
One potential solution is to leverage unlabeled data collected from the end-users' environments.
There is no reliable signal in the target domain to supervise the adaptation process.
We show that this simple additional assumption is sufficient to obtain a potent signal that allows us to perform iterative self-training of 3D object detectors on the target domain.
arXiv Detail & Related papers (2023-03-27T15:07:55Z) - BEV-MAE: Bird's Eye View Masked Autoencoders for Point Cloud
Pre-training in Autonomous Driving Scenarios [51.285561119993105]
We present BEV-MAE, an efficient masked autoencoder pre-training framework for LiDAR-based 3D object detection in autonomous driving.
Specifically, we propose a bird's eye view (BEV) guided masking strategy to guide the 3D encoder learning feature representation.
We introduce a learnable point token to maintain a consistent receptive field size of the 3D encoder.
arXiv Detail & Related papers (2022-12-12T08:15:03Z) - 3D Object Detection with a Self-supervised Lidar Scene Flow Backbone [10.341296683155973]
We propose using a self-supervised training strategy to learn a general point cloud backbone model for downstream 3D vision tasks.
Our main contribution leverages learned flow and motion representations and combines a self-supervised backbone with a 3D detection head.
Experiments on KITTI and nuScenes benchmarks show that the proposed self-supervised pre-training increases 3D detection performance significantly.
arXiv Detail & Related papers (2022-05-02T07:53:29Z) - Attentive Prototypes for Source-free Unsupervised Domain Adaptive 3D
Object Detection [85.11649974840758]
3D object detection networks tend to be biased towards the data they are trained on.
We propose a single-frame approach for source-free, unsupervised domain adaptation of lidar-based 3D object detectors.
arXiv Detail & Related papers (2021-11-30T18:42:42Z) - ST3D++: Denoised Self-training for Unsupervised Domain Adaptation on 3D
Object Detection [78.71826145162092]
We present a self-training method, named ST3D++, with a holistic pseudo label denoising pipeline for unsupervised domain adaptation on 3D object detection.
We equip the pseudo label generation process with a hybrid quality-aware triplet memory to improve the quality and stability of generated pseudo labels.
In the model training stage, we propose a source data assisted training strategy and a curriculum data augmentation policy.
arXiv Detail & Related papers (2021-08-15T07:49:06Z) - ST3D: Self-training for Unsupervised Domain Adaptation on 3D
ObjectDetection [78.71826145162092]
We present a new domain adaptive self-training pipeline, named ST3D, for unsupervised domain adaptation on 3D object detection from point clouds.
Our ST3D achieves state-of-the-art performance on all evaluated datasets and even surpasses fully supervised results on KITTI 3D object detection benchmark.
arXiv Detail & Related papers (2021-03-09T10:51:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.