3D-PL: Domain Adaptive Depth Estimation with 3D-aware Pseudo-Labeling
- URL: http://arxiv.org/abs/2209.09231v1
- Date: Mon, 19 Sep 2022 17:54:17 GMT
- Title: 3D-PL: Domain Adaptive Depth Estimation with 3D-aware Pseudo-Labeling
- Authors: Yu-Ting Yen, Chia-Ni Lu, Wei-Chen Chiu, Yi-Hsuan Tsai
- Abstract summary: We develop a domain adaptation framework via generating reliable pseudo ground truths of depth from real data to provide direct supervisions.
Specifically, we propose two mechanisms for pseudo-labeling: 1) 2D-based pseudo-labels via measuring the consistency of depth predictions when images are with the same content but different styles; 2) 3D-aware pseudo-labels via a point cloud completion network that learns to complete the depth values in the 3D space.
- Score: 37.315964084413174
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: For monocular depth estimation, acquiring ground truths for real data is not
easy, and thus domain adaptation methods are commonly adopted using the
supervised synthetic data. However, this may still incur a large domain gap due
to the lack of supervision from the real data. In this paper, we develop a
domain adaptation framework via generating reliable pseudo ground truths of
depth from real data to provide direct supervisions. Specifically, we propose
two mechanisms for pseudo-labeling: 1) 2D-based pseudo-labels via measuring the
consistency of depth predictions when images are with the same content but
different styles; 2) 3D-aware pseudo-labels via a point cloud completion
network that learns to complete the depth values in the 3D space, thus
providing more structural information in a scene to refine and generate more
reliable pseudo-labels. In experiments, we show that our pseudo-labeling
methods improve depth estimation in various settings, including the usage of
stereo pairs during training. Furthermore, the proposed method performs
favorably against several state-of-the-art unsupervised domain adaptation
approaches in real-world datasets.
Related papers
- Syn-to-Real Unsupervised Domain Adaptation for Indoor 3D Object Detection [50.448520056844885]
We propose a novel framework for syn-to-real unsupervised domain adaptation in indoor 3D object detection.
Our adaptation results from synthetic dataset 3D-FRONT to real-world datasets ScanNetV2 and SUN RGB-D demonstrate remarkable mAP25 improvements of 9.7% and 9.1% over Source-Only baselines.
arXiv Detail & Related papers (2024-06-17T08:18:41Z) - Volumetric Semantically Consistent 3D Panoptic Mapping [77.13446499924977]
We introduce an online 2D-to-3D semantic instance mapping algorithm aimed at generating semantic 3D maps suitable for autonomous agents in unstructured environments.
It introduces novel ways of integrating semantic prediction confidence during mapping, producing semantic and instance-consistent 3D regions.
The proposed method achieves accuracy superior to the state of the art on public large-scale datasets, improving on a number of widely used metrics.
arXiv Detail & Related papers (2023-09-26T08:03:10Z) - Density-Insensitive Unsupervised Domain Adaption on 3D Object Detection [19.703181080679176]
3D object detection from point clouds is crucial in safety-critical autonomous driving.
We propose a density-insensitive domain adaption framework to address the density-induced domain gap.
Experimental results on three widely adopted 3D object detection datasets demonstrate that our proposed domain adaption method outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2023-04-19T06:33:07Z) - Do More With What You Have: Transferring Depth-Scale from Labeled to Unlabeled Domains [43.16293941978469]
Self-supervised depth estimators result in up-to-scale predictions that are linearly correlated to their absolute depth values across the domain.
We show that aligning the field-of-view of two datasets prior to training results in a common linear relationship for both domains.
We use this observed property to transfer the depth-scale from source datasets that have absolute depth labels to new target datasets that lack these measurements.
arXiv Detail & Related papers (2023-03-14T07:07:34Z) - SSDA3D: Semi-supervised Domain Adaptation for 3D Object Detection from
Point Cloud [125.9472454212909]
We present a novel Semi-Supervised Domain Adaptation method for 3D object detection (SSDA3D)
SSDA3D includes an Inter-domain Adaptation stage and an Intra-domain Generalization stage.
Experiments show that, with only 10% labeled target data, our SSDA3D can surpass the fully-supervised oracle model with 100% target label.
arXiv Detail & Related papers (2022-12-06T09:32:44Z) - Geometry-Aware Network for Domain Adaptive Semantic Segmentation [64.00345743710653]
We propose a novel Geometry-Aware Network for Domain Adaptation (GANDA) to shrink the domain gaps.
We exploit 3D topology on the point clouds generated from RGB-D images for coordinate-color disentanglement and pseudo-labels refinement in the target domain.
Our model outperforms state-of-the-arts on GTA5->Cityscapes and SYNTHIA->Cityscapes.
arXiv Detail & Related papers (2022-12-02T00:48:44Z) - ST3D: Self-training for Unsupervised Domain Adaptation on 3D
ObjectDetection [78.71826145162092]
We present a new domain adaptive self-training pipeline, named ST3D, for unsupervised domain adaptation on 3D object detection from point clouds.
Our ST3D achieves state-of-the-art performance on all evaluated datasets and even surpasses fully supervised results on KITTI 3D object detection benchmark.
arXiv Detail & Related papers (2021-03-09T10:51:24Z) - Complete & Label: A Domain Adaptation Approach to Semantic Segmentation
of LiDAR Point Clouds [49.47017280475232]
We study an unsupervised domain adaptation problem for the semantic labeling of 3D point clouds.
We take a Complete and Label approach to recover the underlying surfaces before passing them to a segmentation network.
The recovered 3D surfaces serve as a canonical domain, from which semantic labels can transfer across different LiDAR sensors.
arXiv Detail & Related papers (2020-07-16T17:42:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.