Addressing Data Misalignment in Image-LiDAR Fusion on Point Cloud
Segmentation
- URL: http://arxiv.org/abs/2309.14932v1
- Date: Tue, 26 Sep 2023 13:41:30 GMT
- Title: Addressing Data Misalignment in Image-LiDAR Fusion on Point Cloud
Segmentation
- Authors: Wei Jong Yang, Guan Cheng Lee
- Abstract summary: We show that the projected positions of LiDAR points often misalign on the corresponding image.
In this paper, we would like to address this problem carefully, with a specific focus on the nuScenes dataset and the SOTA of fusion models 2DPASS.
- Score: 0.391609470658968
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the advent of advanced multi-sensor fusion models, there has been a
notable enhancement in the performance of perception tasks within in terms of
autonomous driving. Despite these advancements, the challenges persist,
particularly in the fusion of data from cameras and LiDAR sensors. A critial
concern is the accurate alignment of data from these disparate sensors. Our
observations indicate that the projected positions of LiDAR points often
misalign on the corresponding image. Furthermore, fusion models appear to
struggle in accurately segmenting these misaligned points. In this paper, we
would like to address this problem carefully, with a specific focus on the
nuScenes dataset and the SOTA of fusion models 2DPASS, and providing the
possible solutions or potential improvements.
Related papers
- Refining the ONCE Benchmark with Hyperparameter Tuning [45.55545585587993]
This work focuses on the evaluation of semi-supervised learning approaches for point cloud data.
Data annotation is of paramount importance in the context of LiDAR applications.
We show that improvements from previous semi-supervised methods may not be as profound as previously thought.
arXiv Detail & Related papers (2023-11-10T13:39:07Z) - LiDAR-Camera Panoptic Segmentation via Geometry-Consistent and
Semantic-Aware Alignment [63.83894701779067]
We propose LCPS, the first LiDAR-Camera Panoptic network.
In our approach, we conduct LiDAR-Camera fusion in three stages.
Our fusion strategy improves about 6.9% PQ performance over the LiDAR-only baseline on NuScenes dataset.
arXiv Detail & Related papers (2023-08-03T10:57:58Z) - Towards a Robust Sensor Fusion Step for 3D Object Detection on Corrupted
Data [4.3012765978447565]
This work presents a novel fusion step that addresses data corruptions and makes sensor fusion for 3D object detection more robust.
We demonstrate that our method performs on par with state-of-the-art approaches on normal data and outperforms them on misaligned data.
arXiv Detail & Related papers (2023-06-12T18:06:29Z) - From One to Many: Dynamic Cross Attention Networks for LiDAR and Camera
Fusion [12.792769704561024]
Existing fusion methods tend to align each 3D point to only one projected image pixel based on calibration.
We propose a Dynamic Cross Attention (DCA) module with a novel one-to-many cross-modality mapping.
The whole fusion architecture named Dynamic Cross Attention Network (DCAN) exploits multi-level image features and adapts to multiple representations of point clouds.
arXiv Detail & Related papers (2022-09-25T16:10:14Z) - MSMDFusion: Fusing LiDAR and Camera at Multiple Scales with Multi-Depth
Seeds for 3D Object Detection [89.26380781863665]
Fusing LiDAR and camera information is essential for achieving accurate and reliable 3D object detection in autonomous driving systems.
Recent approaches aim at exploring the semantic densities of camera features through lifting points in 2D camera images into 3D space for fusion.
We propose a novel framework that focuses on the multi-scale progressive interaction of the multi-granularity LiDAR and camera features.
arXiv Detail & Related papers (2022-09-07T12:29:29Z) - Benchmarking the Robustness of LiDAR-Camera Fusion for 3D Object
Detection [58.81316192862618]
Two critical sensors for 3D perception in autonomous driving are the camera and the LiDAR.
fusing these two modalities can significantly boost the performance of 3D perception models.
We benchmark the state-of-the-art fusion methods for the first time.
arXiv Detail & Related papers (2022-05-30T09:35:37Z) - TransFusion: Robust LiDAR-Camera Fusion for 3D Object Detection with
Transformers [49.689566246504356]
We propose TransFusion, a robust solution to LiDAR-camera fusion with a soft-association mechanism to handle inferior image conditions.
TransFusion achieves state-of-the-art performance on large-scale datasets.
We extend the proposed method to the 3D tracking task and achieve the 1st place in the leaderboard of nuScenes tracking.
arXiv Detail & Related papers (2022-03-22T07:15:13Z) - LIF-Seg: LiDAR and Camera Image Fusion for 3D LiDAR Semantic
Segmentation [78.74202673902303]
We propose a coarse-tofine LiDAR and camera fusion-based network (termed as LIF-Seg) for LiDAR segmentation.
The proposed method fully utilizes the contextual information of images and introduces a simple but effective early-fusion strategy.
The cooperation of these two components leads to the success of the effective camera-LiDAR fusion.
arXiv Detail & Related papers (2021-08-17T08:53:11Z) - Deep Learning for Image and Point Cloud Fusion in Autonomous Driving: A
Review [15.10767676137607]
Camera-LiDAR fusion is becoming an emerging research theme.
This paper reviews recent deep-learning-based data fusion approaches that leverage both image and point cloud.
We identify gaps and over-looked challenges between current academic researches and real-world applications.
arXiv Detail & Related papers (2020-04-10T20:43:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.