CO^3: Cooperative Unsupervised 3D Representation Learning for Autonomous
Driving
- URL: http://arxiv.org/abs/2206.04028v1
- Date: Wed, 8 Jun 2022 17:37:58 GMT
- Title: CO^3: Cooperative Unsupervised 3D Representation Learning for Autonomous
Driving
- Authors: Runjian Chen, Yao Mu, Runsen Xu, Wenqi Shao, Chenhan Jiang, Hang Xu,
Zhenguo Li, Ping Luo
- Abstract summary: We propose CO3, namely Cooperative Contrastive Learning and Contextual Shape Prediction, to learn 3D representation for outdoor-scene point clouds in an unsupervised manner.
We believe CO3 will facilitate understanding LiDAR point clouds in outdoor scene.
- Score: 57.16921612272783
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised contrastive learning for indoor-scene point clouds has achieved
great successes. However, unsupervised learning point clouds in outdoor scenes
remains challenging because previous methods need to reconstruct the whole
scene and capture partial views for the contrastive objective. This is
infeasible in outdoor scenes with moving objects, obstacles, and sensors. In
this paper, we propose CO^3, namely Cooperative Contrastive Learning and
Contextual Shape Prediction, to learn 3D representation for outdoor-scene point
clouds in an unsupervised manner. CO^3 has several merits compared to existing
methods. (1) It utilizes LiDAR point clouds from vehicle-side and
infrastructure-side to build views that differ enough but meanwhile maintain
common semantic information for contrastive learning, which are more
appropriate than views built by previous methods. (2) Alongside the contrastive
objective, shape context prediction is proposed as pre-training goal and brings
more task-relevant information for unsupervised 3D point cloud representation
learning, which are beneficial when transferring the learned representation to
downstream detection tasks. (3) As compared to previous methods, representation
learned by CO^3 is able to be transferred to different outdoor scene dataset
collected by different type of LiDAR sensors. (4) CO^3 improves current
state-of-the-art methods on both Once and KITTI datasets by up to 2.58 mAP.
Codes and models will be released. We believe CO^3 will facilitate
understanding LiDAR point clouds in outdoor scene.
Related papers
- GS-PT: Exploiting 3D Gaussian Splatting for Comprehensive Point Cloud Understanding via Self-supervised Learning [15.559369116540097]
Self-supervised learning of point cloud aims to leverage unlabeled 3D data to learn meaningful representations without reliance on manual annotations.
We propose GS-PT, which integrates 3D Gaussian Splatting (3DGS) into point cloud self-supervised learning for the first time.
Our pipeline utilizes transformers as the backbone for self-supervised pre-training and introduces novel contrastive learning tasks through 3DGS.
arXiv Detail & Related papers (2024-09-08T03:46:47Z) - HVDistill: Transferring Knowledge from Images to Point Clouds via Unsupervised Hybrid-View Distillation [106.09886920774002]
We present a hybrid-view-based knowledge distillation framework, termed HVDistill, to guide the feature learning of a point cloud neural network.
Our method achieves consistent improvements over the baseline trained from scratch and significantly out- performs the existing schemes.
arXiv Detail & Related papers (2024-03-18T14:18:08Z) - Self-supervised Learning of LiDAR 3D Point Clouds via 2D-3D Neural Calibration [107.61458720202984]
This paper introduces a novel self-supervised learning framework for enhancing 3D perception in autonomous driving scenes.
We propose the learnable transformation alignment to bridge the domain gap between image and point cloud data.
We establish dense 2D-3D correspondences to estimate the rigid pose.
arXiv Detail & Related papers (2024-01-23T02:41:06Z) - Generalized 3D Self-supervised Learning Framework via Prompted
Foreground-Aware Feature Contrast [38.34558139249363]
We propose a general foreground-aware feature contrast FAC++ framework to learn more effective point cloud representations in pre-training.
We prevent over-discrimination between 3D segments/objects and encourage grouped foreground-to-background distinctions.
We show that our contrast pairs capture clear correspondences among foreground regions during pre-training.
arXiv Detail & Related papers (2023-03-11T11:42:01Z) - CrossPoint: Self-Supervised Cross-Modal Contrastive Learning for 3D
Point Cloud Understanding [2.8661021832561757]
CrossPoint is a simple cross-modal contrastive learning approach to learn transferable 3D point cloud representations.
Our approach outperforms the previous unsupervised learning methods on a diverse range of downstream tasks including 3D object classification and segmentation.
arXiv Detail & Related papers (2022-03-01T18:59:01Z) - Unsupervised Representation Learning for 3D Point Cloud Data [66.92077180228634]
We propose a simple yet effective approach for unsupervised point cloud learning.
In particular, we identify a very useful transformation which generates a good contrastive version of an original point cloud.
We conduct experiments on three downstream tasks which are 3D object classification, shape part segmentation and scene segmentation.
arXiv Detail & Related papers (2021-10-13T10:52:45Z) - Spatio-temporal Self-Supervised Representation Learning for 3D Point
Clouds [96.9027094562957]
We introduce a-temporal representation learning framework, capable of learning from unlabeled tasks.
Inspired by how infants learn from visual data in the wild, we explore rich cues derived from the 3D data.
STRL takes two temporally-related frames from a 3D point cloud sequence as the input, transforms it with the spatial data augmentation, and learns the invariant representation self-supervisedly.
arXiv Detail & Related papers (2021-09-01T04:17:11Z) - PointContrast: Unsupervised Pre-training for 3D Point Cloud
Understanding [107.02479689909164]
In this work, we aim at facilitating research on 3D representation learning.
We measure the effect of unsupervised pre-training on a large source set of 3D scenes.
arXiv Detail & Related papers (2020-07-21T17:59:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.