Cycle and Semantic Consistent Adversarial Domain Adaptation for Reducing
Simulation-to-Real Domain Shift in LiDAR Bird's Eye View
- URL: http://arxiv.org/abs/2104.11021v1
- Date: Thu, 22 Apr 2021 12:47:37 GMT
- Title: Cycle and Semantic Consistent Adversarial Domain Adaptation for Reducing
Simulation-to-Real Domain Shift in LiDAR Bird's Eye View
- Authors: Alejandro Barrera, Jorge Beltr\'an, Carlos Guindel, Jose Antonio
Iglesias, Fernando Garc\'ia
- Abstract summary: We present a BEV domain adaptation method based on CycleGAN that uses prior semantic classification in order to preserve the information of small objects of interest during the domain adaptation process.
The quality of the generated BEVs has been evaluated using a state-of-the-art 3D object detection framework at KITTI 3D Object Detection Benchmark.
- Score: 110.83289076967895
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The performance of object detection methods based on LiDAR information is
heavily impacted by the availability of training data, usually limited to
certain laser devices. As a result, the use of synthetic data is becoming
popular when training neural network models, as both sensor specifications and
driving scenarios can be generated ad-hoc. However, bridging the gap between
virtual and real environments is still an open challenge, as current simulators
cannot completely mimic real LiDAR operation. To tackle this issue, domain
adaptation strategies are usually applied, obtaining remarkable results on
vehicle detection when applied to range view (RV) and bird's eye view (BEV)
projections while failing for smaller road agents. In this paper, we present a
BEV domain adaptation method based on CycleGAN that uses prior semantic
classification in order to preserve the information of small objects of
interest during the domain adaptation process. The quality of the generated
BEVs has been evaluated using a state-of-the-art 3D object detection framework
at KITTI 3D Object Detection Benchmark. The obtained results show the
advantages of the proposed method over the existing alternatives.
Related papers
- CTS: Sim-to-Real Unsupervised Domain Adaptation on 3D Detection [16.96201890965781]
This paper presents a novel framework to transfer models from labeled simulation to unlabeled reality domains.
Experimental results demonstrate that our proposed approach significantly enhances the sim-to-real domain adaptation capability of 3D object detection models.
arXiv Detail & Related papers (2024-06-26T07:31:16Z) - CMDA: Cross-Modal and Domain Adversarial Adaptation for LiDAR-Based 3D
Object Detection [14.063365469339812]
LiDAR-based 3D Object Detection methods often do not generalize well to target domains outside the source (or training) data distribution.
We introduce a novel unsupervised domain adaptation (UDA) method, called CMDA, which leverages visual semantic cues from an image modality.
We also introduce a self-training-based learning strategy, wherein a model is adversarially trained to generate domain-invariant features.
arXiv Detail & Related papers (2024-03-06T14:12:38Z) - Diffusion-Based Particle-DETR for BEV Perception [94.88305708174796]
Bird-Eye-View (BEV) is one of the most widely-used scene representations for visual perception in Autonomous Vehicles (AVs)
Recent diffusion-based methods offer a promising approach to uncertainty modeling for visual perception but fail to effectively detect small objects in the large coverage of the BEV.
Here, we address this problem by combining the diffusion paradigm with current state-of-the-art 3D object detectors in BEV.
arXiv Detail & Related papers (2023-12-18T09:52:14Z) - QE-BEV: Query Evolution for Bird's Eye View Object Detection in Varied Contexts [2.949710700293865]
3D object detection plays a pivotal role in autonomous driving and robotics, demanding precise interpretation of Bird's Eye View (BEV) images.
We introduce a framework utilizing dynamic query evolution strategy, harnesses K-means and Top-K attention mechanisms.
Our evaluation showcases a marked improvement in detection accuracy, setting a new benchmark in the domain of query-based BEV object detection.
arXiv Detail & Related papers (2023-10-07T21:55:29Z) - Unsupervised Domain Adaptation for Self-Driving from Past Traversal
Features [69.47588461101925]
We propose a method to adapt 3D object detectors to new driving environments.
Our approach enhances LiDAR-based detection models using spatial quantized historical features.
Experiments on real-world datasets demonstrate significant improvements.
arXiv Detail & Related papers (2023-09-21T15:00:31Z) - OCBEV: Object-Centric BEV Transformer for Multi-View 3D Object Detection [29.530177591608297]
Multi-view 3D object detection is becoming popular in autonomous driving due to its high effectiveness and low cost.
Most of the current state-of-the-art detectors follow the query-based bird's-eye-view (BEV) paradigm.
We propose an Object-Centric query-BEV detector OCBEV, which can carve the temporal and spatial cues of moving targets more effectively.
arXiv Detail & Related papers (2023-06-02T17:59:48Z) - Adaptive Face Recognition Using Adversarial Information Network [57.29464116557734]
Face recognition models often degenerate when training data are different from testing data.
We propose a novel adversarial information network (AIN) to address it.
arXiv Detail & Related papers (2023-05-23T02:14:11Z) - One-Shot Domain Adaptive and Generalizable Semantic Segmentation with
Class-Aware Cross-Domain Transformers [96.51828911883456]
Unsupervised sim-to-real domain adaptation (UDA) for semantic segmentation aims to improve the real-world test performance of a model trained on simulated data.
Traditional UDA often assumes that there are abundant unlabeled real-world data samples available during training for the adaptation.
We explore the one-shot unsupervised sim-to-real domain adaptation (OSUDA) and generalization problem, where only one real-world data sample is available.
arXiv Detail & Related papers (2022-12-14T15:54:15Z) - Object recognition in atmospheric turbulence scenes [2.657505380055164]
We propose a novel framework that learns distorted features to detect and classify object types in turbulent environments.
Specifically, we utilise deformable convolutions to handle spatial displacement.
We show that the proposed framework outperforms the benchmark with a mean Average Precision (mAP) score exceeding 30%.
arXiv Detail & Related papers (2022-10-25T20:21:25Z) - Attentive Prototypes for Source-free Unsupervised Domain Adaptive 3D
Object Detection [85.11649974840758]
3D object detection networks tend to be biased towards the data they are trained on.
We propose a single-frame approach for source-free, unsupervised domain adaptation of lidar-based 3D object detectors.
arXiv Detail & Related papers (2021-11-30T18:42:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.