Domain generalization of 3D semantic segmentation in autonomous driving
- URL: http://arxiv.org/abs/2212.04245v3
- Date: Thu, 17 Aug 2023 19:15:31 GMT
- Title: Domain generalization of 3D semantic segmentation in autonomous driving
- Authors: Jules Sanchez and Jean-Emmanuel Deschaud and Francois Goulette
- Abstract summary: Despite its importance, domain generalization is relatively unexplored in the case of 3D autonomous driving semantic segmentation.
This paper presents the first benchmark for this application by testing state-of-the-art methods and discussing the difficulty of tackling Laser Imaging Detection and Ranging (LiDAR) domain shifts.
We also propose the first method designed to address this domain generalization, which we call 3DLabelProp.
- Score: 5.240890834159944
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Using deep learning, 3D autonomous driving semantic segmentation has become a
well-studied subject, with methods that can reach very high performance.
Nonetheless, because of the limited size of the training datasets, these models
cannot see every type of object and scene found in real-world applications. The
ability to be reliable in these various unknown environments is called
\textup{domain generalization}.
Despite its importance, domain generalization is relatively unexplored in the
case of 3D autonomous driving semantic segmentation. To fill this gap, this
paper presents the first benchmark for this application by testing
state-of-the-art methods and discussing the difficulty of tackling Laser
Imaging Detection and Ranging (LiDAR) domain shifts.
We also propose the first method designed to address this domain
generalization, which we call 3DLabelProp. This method relies on leveraging the
geometry and sequentiality of the LiDAR data to enhance its generalization
performances by working on partially accumulated point clouds. It reaches a
mean Intersection over Union (mIoU) of 50.4% on SemanticPOSS and of 55.2% on
PandaSet solid-state LiDAR while being trained only on SemanticKITTI, making it
the state-of-the-art method for generalization (+5% and +33% better,
respectively, than the second best method).
The code for this method is available on GitHub:
https://github.com/JulesSanchez/3DLabelProp.
Related papers
- BFANet: Revisiting 3D Semantic Segmentation with Boundary Feature Analysis [33.53327976669034]
We revisit 3D semantic segmentation through a more granular lens, shedding light on subtle complexities that are typically overshadowed by broader performance metrics.
We introduce an innovative 3D semantic segmentation network called BFANet that incorporates detailed analysis of semantic boundary features.
arXiv Detail & Related papers (2025-03-16T15:13:11Z) - Bayesian Self-Training for Semi-Supervised 3D Segmentation [59.544558398992386]
3D segmentation is a core problem in computer vision.
densely labeling 3D point clouds to employ fully-supervised training remains too labor intensive and expensive.
Semi-supervised training provides a more practical alternative, where only a small set of labeled data is given, accompanied by a larger unlabeled set.
arXiv Detail & Related papers (2024-09-12T14:54:31Z) - Augmented Efficiency: Reducing Memory Footprint and Accelerating Inference for 3D Semantic Segmentation through Hybrid Vision [9.96433151449016]
This paper introduces a novel approach to 3D semantic segmentation, distinguished by incorporating a hybrid blend of 2D and 3D computer vision techniques.
We conduct 2D semantic segmentation on RGB images linked to 3D point clouds and extend the results to 3D using an extrusion technique for specific class labels.
This model serves as the current state-of-the-art 3D semantic segmentation model on the KITTI-360 dataset.
arXiv Detail & Related papers (2024-07-23T00:04:10Z) - 3D Adversarial Augmentations for Robust Out-of-Domain Predictions [115.74319739738571]
We focus on improving the generalization to out-of-domain data.
We learn a set of vectors that deform the objects in an adversarial fashion.
We perform adversarial augmentation by applying the learned sample-independent vectors to the available objects when training a model.
arXiv Detail & Related papers (2023-08-29T17:58:55Z) - Weakly Supervised Monocular 3D Object Detection using Multi-View
Projection and Direction Consistency [78.76508318592552]
Monocular 3D object detection has become a mainstream approach in automatic driving for its easy application.
Most current methods still rely on 3D point cloud data for labeling the ground truths used in the training phase.
We propose a new weakly supervised monocular 3D objection detection method, which can train the model with only 2D labels marked on images.
arXiv Detail & Related papers (2023-03-15T15:14:00Z) - Explain What You See: Open-Ended Segmentation and Recognition of
Occluded 3D Objects [0.0]
We propose a novel semantic 3D object-parts segmentation method that has the flexibility of Local-HDP.
We show that the proposed method has a higher percentage of mean intersection over union, using a smaller number of learning instances.
We show that the resulting model produces an explicit set of explanations for the 3D object category recognition task.
arXiv Detail & Related papers (2023-01-17T17:43:46Z) - ALSO: Automotive Lidar Self-supervision by Occupancy estimation [70.70557577874155]
We propose a new self-supervised method for pre-training the backbone of deep perception models operating on point clouds.
The core idea is to train the model on a pretext task which is the reconstruction of the surface on which the 3D points are sampled.
The intuition is that if the network is able to reconstruct the scene surface, given only sparse input points, then it probably also captures some fragments of semantic information.
arXiv Detail & Related papers (2022-12-12T13:10:19Z) - SSDA3D: Semi-supervised Domain Adaptation for 3D Object Detection from
Point Cloud [125.9472454212909]
We present a novel Semi-Supervised Domain Adaptation method for 3D object detection (SSDA3D)
SSDA3D includes an Inter-domain Adaptation stage and an Intra-domain Generalization stage.
Experiments show that, with only 10% labeled target data, our SSDA3D can surpass the fully-supervised oracle model with 100% target label.
arXiv Detail & Related papers (2022-12-06T09:32:44Z) - PerMO: Perceiving More at Once from a Single Image for Autonomous
Driving [76.35684439949094]
We present a novel approach to detect, segment, and reconstruct complete textured 3D models of vehicles from a single image.
Our approach combines the strengths of deep learning and the elegance of traditional techniques.
We have integrated these algorithms with an autonomous driving system.
arXiv Detail & Related papers (2020-07-16T05:02:45Z) - 3D-MiniNet: Learning a 2D Representation from Point Clouds for Fast and
Efficient 3D LIDAR Semantic Segmentation [9.581605678437032]
3D-MiniNet is a novel approach for LIDAR semantic segmentation that combines 3D and 2D learning layers.
It first learns a 2D representation from the raw points through a novel projection which extracts local and global information from the 3D data.
These 2D semantic labels are re-projected back to the 3D space and enhanced through a post-processing module.
arXiv Detail & Related papers (2020-02-25T14:33:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.