Deformation and Correspondence Aware Unsupervised Synthetic-to-Real
Scene Flow Estimation for Point Clouds
- URL: http://arxiv.org/abs/2203.16895v1
- Date: Thu, 31 Mar 2022 09:03:23 GMT
- Title: Deformation and Correspondence Aware Unsupervised Synthetic-to-Real
Scene Flow Estimation for Point Clouds
- Authors: Zhao Jin, Yinjie Lei, Naveed Akhtar, Haifeng Li, Munawar Hayat
- Abstract summary: We develop a point cloud collector and scene flow annotator for GTA-V engine to automatically obtain diverse training samples without human intervention.
We propose a mean-teacher-based domain adaptation framework that leverages self-generated pseudo-labels of the target domain.
Our framework achieves superior adaptation performance on six source-target dataset pairs, remarkably closing the average domain gap by 60%.
- Score: 43.792032657561236
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Point cloud scene flow estimation is of practical importance for dynamic
scene navigation in autonomous driving. Since scene flow labels are hard to
obtain, current methods train their models on synthetic data and transfer them
to real scenes. However, large disparities between existing synthetic datasets
and real scenes lead to poor model transfer. We make two major contributions to
address that. First, we develop a point cloud collector and scene flow
annotator for GTA-V engine to automatically obtain diverse realistic training
samples without human intervention. With that, we develop a large-scale
synthetic scene flow dataset GTA-SF. Second, we propose a mean-teacher-based
domain adaptation framework that leverages self-generated pseudo-labels of the
target domain. It also explicitly incorporates shape deformation regularization
and surface correspondence refinement to address distortions and misalignments
in domain transfer. Through extensive experiments, we show that our GTA-SF
dataset leads to a consistent boost in model generalization to three real
datasets (i.e., Waymo, Lyft and KITTI) as compared to the most widely used FT3D
dataset. Moreover, our framework achieves superior adaptation performance on
six source-target dataset pairs, remarkably closing the average domain gap by
60%. Data and codes are available at https://github.com/leolyj/DCA-SRSFE
Related papers
- A New Benchmark: On the Utility of Synthetic Data with Blender for Bare
Supervised Learning and Downstream Domain Adaptation [42.2398858786125]
Deep learning in computer vision has achieved great success with the price of large-scale labeled training data.
The uncontrollable data collection process produces non-IID training and test data, where undesired duplication may exist.
To circumvent them, an alternative is to generate synthetic data via 3D rendering with domain randomization.
arXiv Detail & Related papers (2023-03-16T09:03:52Z) - One-Shot Domain Adaptive and Generalizable Semantic Segmentation with
Class-Aware Cross-Domain Transformers [96.51828911883456]
Unsupervised sim-to-real domain adaptation (UDA) for semantic segmentation aims to improve the real-world test performance of a model trained on simulated data.
Traditional UDA often assumes that there are abundant unlabeled real-world data samples available during training for the adaptation.
We explore the one-shot unsupervised sim-to-real domain adaptation (OSUDA) and generalization problem, where only one real-world data sample is available.
arXiv Detail & Related papers (2022-12-14T15:54:15Z) - TRoVE: Transforming Road Scene Datasets into Photorealistic Virtual
Environments [84.6017003787244]
This work proposes a synthetic data generation pipeline to address the difficulties and domain-gaps present in simulated datasets.
We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation.
arXiv Detail & Related papers (2022-08-16T20:46:08Z) - GIPSO: Geometrically Informed Propagation for Online Adaptation in 3D
LiDAR Segmentation [60.07812405063708]
3D point cloud semantic segmentation is fundamental for autonomous driving.
Most approaches in the literature neglect an important aspect, i.e., how to deal with domain shift when handling dynamic scenes.
This paper advances the state of the art in this research field.
arXiv Detail & Related papers (2022-07-20T09:06:07Z) - Towards 3D Scene Understanding by Referring Synthetic Models [65.74211112607315]
Methods typically alleviate on-extensive annotations on real scene scans.
We explore how synthetic models rely on real scene categories of synthetic features to a unified feature space.
Experiments show that our method achieves the average mAP of 46.08% on the ScanNet S3DIS dataset and 55.49% by learning datasets.
arXiv Detail & Related papers (2022-03-20T13:06:15Z) - Scalable Scene Flow from Point Clouds in the Real World [30.437100097997245]
We introduce a new large scale benchmark for scene flow based on the Open dataset.
We show how previous works were bounded based on the amount of real LiDAR data available.
We introduce the model architecture FastFlow3D that provides real time inference on the full point cloud.
arXiv Detail & Related papers (2021-03-01T20:56:05Z) - ePointDA: An End-to-End Simulation-to-Real Domain Adaptation Framework
for LiDAR Point Cloud Segmentation [111.56730703473411]
Training deep neural networks (DNNs) on LiDAR data requires large-scale point-wise annotations.
Simulation-to-real domain adaptation (SRDA) trains a DNN using unlimited synthetic data with automatically generated labels.
ePointDA consists of three modules: self-supervised dropout noise rendering, statistics-invariant and spatially-adaptive feature alignment, and transferable segmentation learning.
arXiv Detail & Related papers (2020-09-07T23:46:08Z) - Virtual to Real adaptation of Pedestrian Detectors [9.432150710329607]
ViPeD is a new synthetically generated set of images collected with the graphical engine of the video game GTA V - Grand Theft Auto V.
We propose two different Domain Adaptation techniques suitable for the pedestrian detection task, but possibly applicable to general object detection.
Experiments show that the network trained with ViPeD can generalize over unseen real-world scenarios better than the detector trained over real-world data.
arXiv Detail & Related papers (2020-01-09T14:50:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.