Walking Your LiDOG: A Journey Through Multiple Domains for LiDAR
Semantic Segmentation
- URL: http://arxiv.org/abs/2304.11705v2
- Date: Tue, 29 Aug 2023 10:08:24 GMT
- Title: Walking Your LiDOG: A Journey Through Multiple Domains for LiDAR
Semantic Segmentation
- Authors: Cristiano Saltori and Aljo\v{s}a O\v{s}ep and Elisa Ricci and Laura
Leal-Taix\'e
- Abstract summary: We study domain generalization for LiDAR semantic segmentation (DG-LSS)
Our results confirm a significant gap between methods, evaluated in a cross-domain setting.
We propose the first method specifically designed for DG-LSS, which obtains $34.88$ mIoU on the target domain.
- Score: 16.857861952638498
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The ability to deploy robots that can operate safely in diverse environments
is crucial for developing embodied intelligent agents. As a community, we have
made tremendous progress in within-domain LiDAR semantic segmentation. However,
do these methods generalize across domains? To answer this question, we design
the first experimental setup for studying domain generalization (DG) for LiDAR
semantic segmentation (DG-LSS). Our results confirm a significant gap between
methods, evaluated in a cross-domain setting: for example, a model trained on
the source dataset (SemanticKITTI) obtains $26.53$ mIoU on the target data,
compared to $48.49$ mIoU obtained by the model trained on the target domain
(nuScenes). To tackle this gap, we propose the first method specifically
designed for DG-LSS, which obtains $34.88$ mIoU on the target domain,
outperforming all baselines. Our method augments a sparse-convolutional
encoder-decoder 3D segmentation network with an additional, dense 2D
convolutional decoder that learns to classify a birds-eye view of the point
cloud. This simple auxiliary task encourages the 3D network to learn features
that are robust to sensor placement shifts and resolution, and are transferable
across domains. With this work, we aim to inspire the community to develop and
evaluate future models in such cross-domain conditions.
Related papers
- CMDA: Cross-Modal and Domain Adversarial Adaptation for LiDAR-Based 3D
Object Detection [14.063365469339812]
LiDAR-based 3D Object Detection methods often do not generalize well to target domains outside the source (or training) data distribution.
We introduce a novel unsupervised domain adaptation (UDA) method, called CMDA, which leverages visual semantic cues from an image modality.
We also introduce a self-training-based learning strategy, wherein a model is adversarially trained to generate domain-invariant features.
arXiv Detail & Related papers (2024-03-06T14:12:38Z) - Domain Generalization of 3D Object Detection by Density-Resampling [14.510085711178217]
Point-cloud-based 3D object detection suffers from performance degradation when encountering data with novel domain gaps.
We propose an SDG method to improve the generalizability of 3D object detection to unseen target domains.
Our work introduces a novel data augmentation method and contributes a new multi-task learning strategy in the methodology.
arXiv Detail & Related papers (2023-11-17T20:01:29Z) - Compositional Semantic Mix for Domain Adaptation in Point Cloud
Segmentation [65.78246406460305]
compositional semantic mixing represents the first unsupervised domain adaptation technique for point cloud segmentation.
We present a two-branch symmetric network architecture capable of concurrently processing point clouds from a source domain (e.g. synthetic) and point clouds from a target domain (e.g. real-world)
arXiv Detail & Related papers (2023-08-28T14:43:36Z) - BEV-DG: Cross-Modal Learning under Bird's-Eye View for Domain
Generalization of 3D Semantic Segmentation [59.99683295806698]
Cross-modal Unsupervised Domain Adaptation (UDA) aims to exploit the complementarity of 2D-3D data to overcome the lack of annotation in a new domain.
We propose cross-modal learning under bird's-eye view for Domain Generalization (DG) of 3D semantic segmentation, called BEV-DG.
arXiv Detail & Related papers (2023-08-12T11:09:17Z) - Bi3D: Bi-domain Active Learning for Cross-domain 3D Object Detection [32.29833072399945]
We propose a Bi-domain active learning approach, namely Bi3D, to solve the cross-domain 3D object detection task.
Bi3D achieves a promising target-domain detection accuracy (89.63% on KITTI) compared with UDAbased work (84.29%), even surpassing the detector trained on the full set of the labeled target domain.
arXiv Detail & Related papers (2023-03-10T12:38:37Z) - SSDA3D: Semi-supervised Domain Adaptation for 3D Object Detection from
Point Cloud [125.9472454212909]
We present a novel Semi-Supervised Domain Adaptation method for 3D object detection (SSDA3D)
SSDA3D includes an Inter-domain Adaptation stage and an Intra-domain Generalization stage.
Experiments show that, with only 10% labeled target data, our SSDA3D can surpass the fully-supervised oracle model with 100% target label.
arXiv Detail & Related papers (2022-12-06T09:32:44Z) - Geometry-Aware Network for Domain Adaptive Semantic Segmentation [64.00345743710653]
We propose a novel Geometry-Aware Network for Domain Adaptation (GANDA) to shrink the domain gaps.
We exploit 3D topology on the point clouds generated from RGB-D images for coordinate-color disentanglement and pseudo-labels refinement in the target domain.
Our model outperforms state-of-the-arts on GTA5->Cityscapes and SYNTHIA->Cityscapes.
arXiv Detail & Related papers (2022-12-02T00:48:44Z) - Stagewise Unsupervised Domain Adaptation with Adversarial Self-Training
for Road Segmentation of Remote Sensing Images [93.50240389540252]
Road segmentation from remote sensing images is a challenging task with wide ranges of application potentials.
We propose a novel stagewise domain adaptation model called RoadDA to address the domain shift (DS) issue in this field.
Experiment results on two benchmarks demonstrate that RoadDA can efficiently reduce the domain gap and outperforms state-of-the-art methods.
arXiv Detail & Related papers (2021-08-28T09:29:14Z) - Unsupervised Domain Adaptive 3D Detection with Multi-Level Consistency [90.71745178767203]
Deep learning-based 3D object detection has achieved unprecedented success with the advent of large-scale autonomous driving datasets.
Existing 3D domain adaptive detection methods often assume prior access to the target domain annotations, which is rarely feasible in the real world.
We study a more realistic setting, unsupervised 3D domain adaptive detection, which only utilizes source domain annotations.
arXiv Detail & Related papers (2021-07-23T17:19:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.