Multi-view Structural Convolution Network for Domain-Invariant Point Cloud Recognition of Autonomous Vehicles
- URL: http://arxiv.org/abs/2501.16289v2
- Date: Wed, 26 Feb 2025 15:53:53 GMT
- Title: Multi-view Structural Convolution Network for Domain-Invariant Point Cloud Recognition of Autonomous Vehicles
- Authors: Younggun Kim, Beomsik Cho, Seonghoon Ryoo, Soomok Lee,
- Abstract summary: Multi-View Structural Convolution Network (MSCN) designed for domain-invariant point cloud recognition.<n> MSCN comprises Structural Convolution Layers (SCL) that extract local context geometric features from point clouds.<n> MSCN enhances feature representation by training with unseen domain point clouds derived from source domain point clouds.
- Score: 3.3748750222488657
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Point cloud representation has recently become a research hotspot in the field of computer vision and has been utilized for autonomous vehicles. However, adapting deep learning networks for point cloud data recognition is challenging due to the variability in datasets and sensor technologies. This variability underscores the necessity for adaptive techniques to maintain accuracy under different conditions. In this paper, we present the Multi-View Structural Convolution Network (MSCN) designed for domain-invariant point cloud recognition. MSCN comprises Structural Convolution Layers (SCL) that extract local context geometric features from point clouds and Structural Aggregation Layers (SAL) that extract and aggregate both local and overall context features from point clouds. Additionally, our MSCN enhances feature representation robustness by training with unseen domain point clouds derived from source domain point clouds. This method acquires domain-invariant features and exhibits robust, consistent performance across various point cloud datasets, ensuring compatibility with diverse sensor configurations without the need for parameter adjustments. This highlights MSCN's potential to significantly improve the reliability and domain invariant features in different environments. Our code is available at https://github.com/MLMLab/MSCN.
Related papers
- High-Quality Pseudo-Label Generation Based on Visual Prompt Assisted Cloud Model Update [6.11503045313947]
Existing methods often assume reliable cloud models, neglecting potential errors or struggling with complex distribution shifts.
This paper proposes Cloud-Adaptive High-Quality Pseudo-label generation (CA-HQP)
Experiments on the Bellevue traffic dataset demonstrate that CA-HQP significantly improves pseudo-label quality compared to existing methods.
arXiv Detail & Related papers (2025-04-01T08:20:16Z) - PointDGMamba: Domain Generalization of Point Cloud Classification via Generalized State Space Model [77.00221501105788]
Domain Generalization (DG) has been recently explored to improve the generalizability of point cloud classification.<n>We present the first work that studies the generalizability of state space models (SSMs) in DG PCC.<n>We propose a novel framework, PointDGMamba, that excels in strong generalizability toward unseen domains.
arXiv Detail & Related papers (2024-08-24T12:53:48Z) - 3D Adaptive Structural Convolution Network for Domain-Invariant Point Cloud Recognition [3.3748750222488657]
3D Adaptive Structural Convolution Network (3D-ASCN) is a cutting-edge framework for 3D point cloud recognition.
It combines 3D convolution kernels, a structural tree structure, and adaptive neighborhood sampling for effective geometric feature extraction.
arXiv Detail & Related papers (2024-07-05T19:38:10Z) - Point Cloud Compression with Implicit Neural Representations: A Unified Framework [54.119415852585306]
We present a pioneering point cloud compression framework capable of handling both geometry and attribute components.
Our framework utilizes two coordinate-based neural networks to implicitly represent a voxelized point cloud.
Our method exhibits high universality when contrasted with existing learning-based techniques.
arXiv Detail & Related papers (2024-05-19T09:19:40Z) - InvariantOODG: Learning Invariant Features of Point Clouds for
Out-of-Distribution Generalization [17.96808017359983]
We propose InvariantOODG, which learns invariability between point clouds with different distributions.
We define a set of learnable anchor points that locate the most useful local regions and two types of transformations to augment the input point clouds.
The experimental results demonstrate the effectiveness of the proposed model on 3D domain generalization benchmarks.
arXiv Detail & Related papers (2024-01-08T09:41:22Z) - Compositional Semantic Mix for Domain Adaptation in Point Cloud
Segmentation [65.78246406460305]
compositional semantic mixing represents the first unsupervised domain adaptation technique for point cloud segmentation.
We present a two-branch symmetric network architecture capable of concurrently processing point clouds from a source domain (e.g. synthetic) and point clouds from a target domain (e.g. real-world)
arXiv Detail & Related papers (2023-08-28T14:43:36Z) - Collect-and-Distribute Transformer for 3D Point Cloud Analysis [82.03517861433849]
We propose a new transformer network equipped with a collect-and-distribute mechanism to communicate short- and long-range contexts of point clouds.
Results show the effectiveness of the proposed CDFormer, delivering several new state-of-the-art performances on point cloud classification and segmentation tasks.
arXiv Detail & Related papers (2023-06-02T03:48:45Z) - Adaptive Edge-to-Edge Interaction Learning for Point Cloud Analysis [118.30840667784206]
Key issue for point cloud data processing is extracting useful information from local regions.
Previous works ignore the relation between edges in local regions, which encodes the local shape information.
This paper proposes a novel Adaptive Edge-to-Edge Interaction Learning module.
arXiv Detail & Related papers (2022-11-20T07:10:14Z) - ePointDA: An End-to-End Simulation-to-Real Domain Adaptation Framework
for LiDAR Point Cloud Segmentation [111.56730703473411]
Training deep neural networks (DNNs) on LiDAR data requires large-scale point-wise annotations.
Simulation-to-real domain adaptation (SRDA) trains a DNN using unlimited synthetic data with automatically generated labels.
ePointDA consists of three modules: self-supervised dropout noise rendering, statistics-invariant and spatially-adaptive feature alignment, and transferable segmentation learning.
arXiv Detail & Related papers (2020-09-07T23:46:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.