SoftPool++: An Encoder-Decoder Network for Point Cloud Completion
- URL: http://arxiv.org/abs/2205.03899v1
- Date: Sun, 8 May 2022 15:31:36 GMT
- Title: SoftPool++: An Encoder-Decoder Network for Point Cloud Completion
- Authors: Yida Wang, David Joseph Tan, Nassir Navab, Federico Tombari
- Abstract summary: We propose a novel convolutional operator for the task of point cloud completion.
The proposed operator does not require any max-pooling or voxelization operation.
We show that our approach achieves state-of-the-art performance in shape completion at low and high resolutions.
- Score: 93.54286830844134
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel convolutional operator for the task of point cloud
completion. One striking characteristic of our approach is that, conversely to
related work it does not require any max-pooling or voxelization operation.
Instead, the proposed operator used to learn the point cloud embedding in the
encoder extracts permutation-invariant features from the point cloud via a
soft-pooling of feature activations, which are able to preserve fine-grained
geometric details. These features are then passed on to a decoder architecture.
Due to the compression in the encoder, a typical limitation of this type of
architectures is that they tend to lose parts of the input shape structure. We
propose to overcome this limitation by using skip connections specifically
devised for point clouds, where links between corresponding layers in the
encoder and the decoder are established. As part of these connections, we
introduce a transformation matrix that projects the features from the encoder
to the decoder and vice-versa. The quantitative and qualitative results on the
task of object completion from partial scans on the ShapeNet dataset show that
incorporating our approach achieves state-of-the-art performance in shape
completion both at low and high resolutions.
Related papers
- Distributed Semantic Segmentation with Efficient Joint Source and Task Decoding [18.8803233019656]
Deep neural networks (DNNs) execute one part of the network on edge devices and the other part on a large-scale cloud platform.
In this work, we propose joint source and task decoding, as it allows for a smaller network size in the cloud.
We demonstrate the effectiveness of our method by achieving a distributed semantic segmentation SOTA over a wide range of intersections.
arXiv Detail & Related papers (2024-07-15T20:20:04Z) - Lightweight super resolution network for point cloud geometry
compression [34.42460388539782]
We present an approach for compressing point cloud geometry by leveraging a lightweight super-resolution network.
The proposed method involves decomposing a point cloud into a base point cloud and the patterns for reconstructing the original point cloud.
Experiments on MPEG Cat1 (Solid) and Cat2 datasets demonstrate the remarkable compression performance achieved by our method.
arXiv Detail & Related papers (2023-11-02T03:34:51Z) - EPCL: Frozen CLIP Transformer is An Efficient Point Cloud Encoder [60.52613206271329]
This paper introduces textbfEfficient textbfPoint textbfCloud textbfLearning (EPCL) for training high-quality point cloud models with a frozen CLIP transformer.
Our EPCL connects the 2D and 3D modalities by semantically aligning the image features and point cloud features without paired 2D-3D data.
arXiv Detail & Related papers (2022-12-08T06:27:11Z) - SeRP: Self-Supervised Representation Learning Using Perturbed Point
Clouds [6.29475963948119]
SeRP consists of encoder-decoder architecture that takes perturbed or corrupted point clouds as inputs.
We have used Transformers and PointNet-based Autoencoders.
arXiv Detail & Related papers (2022-09-13T15:22:36Z) - Learning Local Displacements for Point Cloud Completion [93.54286830844134]
We propose a novel approach aimed at object and semantic scene completion from a partial scan represented as a 3D point cloud.
Our architecture relies on three novel layers that are used successively within an encoder-decoder structure.
We evaluate both architectures on object and indoor scene completion tasks, achieving state-of-the-art performance.
arXiv Detail & Related papers (2022-03-30T18:31:37Z) - Upsampling Autoencoder for Self-Supervised Point Cloud Learning [11.19408173558718]
We propose a self-supervised pretraining model for point cloud learning without human annotations.
Upsampling operation encourages the network to capture both high-level semantic information and low-level geometric information of the point cloud.
We find that our UAE outperforms previous state-of-the-art methods in shape classification, part segmentation and point cloud upsampling tasks.
arXiv Detail & Related papers (2022-03-21T07:20:37Z) - PoinTr: Diverse Point Cloud Completion with Geometry-Aware Transformers [81.71904691925428]
We present a new method that reformulates point cloud completion as a set-to-set translation problem.
We also design a new model, called PoinTr, that adopts a transformer encoder-decoder architecture for point cloud completion.
Our method outperforms state-of-the-art methods by a large margin on both the new benchmarks and the existing ones.
arXiv Detail & Related papers (2021-08-19T17:58:56Z) - Dynamic Neural Representational Decoders for High-Resolution Semantic
Segmentation [98.05643473345474]
We propose a novel decoder, termed dynamic neural representational decoder (NRD)
As each location on the encoder's output corresponds to a local patch of the semantic labels, in this work, we represent these local patches of labels with compact neural networks.
This neural representation enables our decoder to leverage the smoothness prior in the semantic label space, and thus makes our decoder more efficient.
arXiv Detail & Related papers (2021-07-30T04:50:56Z) - Beyond Single Stage Encoder-Decoder Networks: Deep Decoders for Semantic
Image Segmentation [56.44853893149365]
Single encoder-decoder methodologies for semantic segmentation are reaching their peak in terms of segmentation quality and efficiency per number of layers.
We propose a new architecture based on a decoder which uses a set of shallow networks for capturing more information content.
In order to further improve the architecture we introduce a weight function which aims to re-balance classes to increase the attention of the networks to under-represented objects.
arXiv Detail & Related papers (2020-07-19T18:44:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.