Leveraging Single-View Images for Unsupervised 3D Point Cloud Completion
- URL: http://arxiv.org/abs/2212.00564v3
- Date: Wed, 6 Dec 2023 02:10:09 GMT
- Title: Leveraging Single-View Images for Unsupervised 3D Point Cloud Completion
- Authors: Lintai Wu, Qijian Zhang, Junhui Hou, and Yong Xu
- Abstract summary: Cross-PCC is an unsupervised point cloud completion method without requiring any 3D complete point clouds.
To take advantage of the complementary information from 2D images, we use a single-view RGB image to extract 2D features.
Our method even achieves comparable performance to some supervised methods.
- Score: 53.93172686610741
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Point clouds captured by scanning devices are often incomplete due to
occlusion. To overcome this limitation, point cloud completion methods have
been developed to predict the complete shape of an object based on its partial
input. These methods can be broadly classified as supervised or unsupervised.
However, both categories require a large number of 3D complete point clouds,
which may be difficult to capture. In this paper, we propose Cross-PCC, an
unsupervised point cloud completion method without requiring any 3D complete
point clouds. We only utilize 2D images of the complete objects, which are
easier to capture than 3D complete and clean point clouds. Specifically, to
take advantage of the complementary information from 2D images, we use a
single-view RGB image to extract 2D features and design a fusion module to fuse
the 2D and 3D features extracted from the partial point cloud. To guide the
shape of predicted point clouds, we project the predicted points of the object
to the 2D plane and use the foreground pixels of its silhouette maps to
constrain the position of the projected points. To reduce the outliers of the
predicted point clouds, we propose a view calibrator to move the points
projected to the background into the foreground by the single-view silhouette
image. To the best of our knowledge, our approach is the first point cloud
completion method that does not require any 3D supervision. The experimental
results of our method are superior to those of the state-of-the-art
unsupervised methods by a large margin. Moreover, our method even achieves
comparable performance to some supervised methods. We will make the source code
publicly available at https://github.com/ltwu6/cross-pcc.
Related papers
- Adapt PointFormer: 3D Point Cloud Analysis via Adapting 2D Visual Transformers [38.08724410736292]
This paper attempts to leverage pre-trained models with 2D prior knowledge to accomplish the tasks for 3D point cloud analysis.
We propose the Adaptive PointFormer (APF), which fine-tunes pre-trained 2D models with only a modest number of parameters to directly process point clouds.
arXiv Detail & Related papers (2024-07-18T06:32:45Z) - Self-supervised 3D Point Cloud Completion via Multi-view Adversarial Learning [61.14132533712537]
We propose MAL-SPC, a framework that effectively leverages both object-level and category-specific geometric similarities to complete missing structures.
Our MAL-SPC does not require any 3D complete supervision and only necessitates a single partial point cloud for each object.
arXiv Detail & Related papers (2024-07-13T06:53:39Z) - Intrinsic Image Decomposition Using Point Cloud Representation [13.771632868567277]
We introduce Point Intrinsic Net (PoInt-Net), which leverages 3D point cloud data to concurrently estimate albedo and shading maps.
PoInt-Net is efficient, achieving consistent performance across point clouds of any size with training only required on small-scale point clouds.
arXiv Detail & Related papers (2023-07-20T14:51:28Z) - CrossPoint: Self-Supervised Cross-Modal Contrastive Learning for 3D
Point Cloud Understanding [2.8661021832561757]
CrossPoint is a simple cross-modal contrastive learning approach to learn transferable 3D point cloud representations.
Our approach outperforms the previous unsupervised learning methods on a diverse range of downstream tasks including 3D object classification and segmentation.
arXiv Detail & Related papers (2022-03-01T18:59:01Z) - Unsupervised Learning of Fine Structure Generation for 3D Point Clouds
by 2D Projection Matching [66.98712589559028]
We propose an unsupervised approach for 3D point cloud generation with fine structures.
Our method can recover fine 3D structures from 2D silhouette images at different resolutions.
arXiv Detail & Related papers (2021-08-08T22:15:31Z) - ParaNet: Deep Regular Representation for 3D Point Clouds [62.81379889095186]
ParaNet is a novel end-to-end deep learning framework for representing 3D point clouds.
It converts an irregular 3D point cloud into a regular 2D color image, named point geometry image (PGI)
In contrast to conventional regular representation modalities based on multi-view projection and voxelization, the proposed representation is differentiable and reversible.
arXiv Detail & Related papers (2020-12-05T13:19:55Z) - From Image Collections to Point Clouds with Self-supervised Shape and
Pose Networks [53.71440550507745]
Reconstructing 3D models from 2D images is one of the fundamental problems in computer vision.
We propose a deep learning technique for 3D object reconstruction from a single image.
We learn both 3D point cloud reconstruction and pose estimation networks in a self-supervised manner.
arXiv Detail & Related papers (2020-05-05T04:25:16Z) - ImVoteNet: Boosting 3D Object Detection in Point Clouds with Image Votes [93.82668222075128]
We propose a 3D detection architecture called ImVoteNet for RGB-D scenes.
ImVoteNet is based on fusing 2D votes in images and 3D votes in point clouds.
We validate our model on the challenging SUN RGB-D dataset.
arXiv Detail & Related papers (2020-01-29T05:09:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.